@movq@www.uninformativ.de WE NEED MORE BACKUPS!!!!!!!!!!!1
@kat@yarn.girlonthemoon.xyz Oh no. šØ Backups! We need more backups!
linodeās having a major outage (ongoing as of writing, over 24 hours in) and my friend runs a site i help out with on one of their servers. we didnāt have recent backups so i got really anxious about possible severe data loss considering the situation with linode doesnāt look great (it seems like a really bad incident).
ā¦anyway the server magically came back online and i got backups of the whole application and database, iām so relieved :ā)
Running monthly backupsā¦
@xuu@txt.sour.is or @kat@yarn.girlonthemoon.xyz Do either of you have time this weekend to test upgrading your pod to the new cacher
branch? š¤ It is recommended you take a full backup of you pod beforehand, just in case. Keen to get this branch merged and to cut a new release finally after >2 years š¤£
@bender@twtxt.net Lemme look at the old backupā¦
@prologic@twtxt.net @bmallred@staystrong.run So is restic considered stable by now? āStableā as in āstable data formatā, like a future version will still be able to retrieve my current backups. I mean, itās at version ā0.18ā, but they donāt specify which versioning scheme they use.
I use restic and Backblaze B2 for offline backup storage at a cost of $6/TB/month. I donāt backup my entire ~20TB NAS and its datasets however, so Iām only paying about ~$2/month right now. I only backup the most important things I cannot afford to lose or annot re-created.
@movq@www.uninformativ.de there are many other similar backup tools. I would love to hear what will make you pick Borg above the rest.
On top of my usual backups (which are already offsite, but it requires me carrying a hard disk to that other site), I think I might rent a storage server and use Borg. š¤ Hoping that their encryption is good enough. Maybe thatāll also finally convince me to get a faster internet connection. š
Add support for skipping backup if data is unchagned Ā· 0cf9514e9e - backup-docker-volumes - Mills š I just discovered today, when running backups, that this commit is why my backups stopped working for the last 4 months. It wasnāt that I was forgetting to do them every month, I broke the fucking tool 𤣠Fuck š¤¦āāļø
@thecanine@twtxt.net I mean I can restore whatever anyone likes, the problem is the last backup I took was 4 months ago š So I decided to start over (from scratch). Just let me know what you want and Iāll do it! I used the 4-month old backup to restore your account (by hand) and avatar at least š¤£
@thecanine@twtxt.net Iām so sorry I fucked things up š„² I hope you can trust Iāll try to do a better job of backups and data going forward š¤
This weekend (as some of you may now) I accidently nuke this Podās entire data volume š¤¦āāļø What a disastrous incident 𤣠I decided instead of trying to restore from a 4-month old backup (weāll get into why I hadnāt been taking backups consistently later), that weād start a fresh! š Spring clean! š§¼ ā Anyway⦠One of the things I realised was I was missing a very critical Safety Controls in my own ways of working⦠Iāve now rectified thisā¦
@prologic@twtxt.net Spring cleanup! Thatās one way to encourage people to self-host their feeds. :-D
Since Iām only interested in the url
metadata field for hashing, I do not keep any comments or metadata for that matter, just the messages themselves. The last time I fetched was probably some time yesterday evening (UTC+2). I cannot tell exactly, because the recorded last fetch timestamp has been overridden with todayās by now.
I dumped my new SQLite cache into: https://lyse.isobeef.org/tmp/backup.tar.gz This time maybe even correctly, if youāre lucky. Iām not entirely sure. It took me a few attempts (date and time were separated by space instead of T
at first, I normalized offsets +00:00
to Z
as yarnd does and converted newlines back to U+2028
). At least now the simple cross check with the Twtxt Feed Validator does not yield any problems.
Oh well. Iāve gone and done it again! This time Iāve lost 4 months of data because for some reason Iāve been busy and havenāt been taking backups of all the things I should be?! š¤ Farrrrk š¤¬
@prologic@twtxt.net If it develops, and Iām not saying it will happen soon, perhaps Yarn could be connected as an additional node. Implementation would not be difficult for any client or software. It will not only be a backup of twtxt, but it will be the source for search, discovery and network health.
i upgraded my pc from lubuntu 22.04 to 24.04 yesterday and i was like āsurely there is no way this will go smoothlyā but no it somehow did. like i didnāt take a backup i just said fuck it and upgraded and it WORKED?!?! i mean i had some driver issues but it wasnāt too bad to fix. wild
A random suggestion. You should add a password to your private ssh key. Why? If someone steals your key, they wonāt be able to do anything without the password.
You should run: ssh-keygen -p
And remember to make a backup copy of key file. As a developer, it is a one of the most valuable files on your computer.
@emmanuel@wald.ovh Btw I already figured out why accessing your web server is slow:
$ host wald.ovh
wald.ovh has address 86.243.228.45
wald.ovh has address 90.19.202.229
wald.ovh
has 2 IPv4 addresses, one of which is dead and doesnāt respond.. Thatās why accessing your website is so slow as depending on client and browser behaviors one of two things may happen 1) a random IP is chosen and ½ the time the wrong one is picked or 2) both are tried in some random order and ½ the time its slow because the broken one is picked.
If you donāt know what 86.243.228.45
is, or itās a dead backup server or something, Iād suggest you remove this from the domain record.
Iāve talked about how I do backups on unix a bunch of times, but someone asked again today and I realized I didnāt have it written down where I could point to. So I wrote a lab report: http://a.9srv.net/reports/index.html#vac-unix
Can I get someone like maybe @xuu@txt.sour.is or @abucci@anthony.buc.ci or even @eldersnake@we.loveprivacy.club ā If you have some spare time ā to test this yarnd
PR that upgrades the Bitcask dependency for its internal database to v2? š
VERY IMPORTANT If you do; Please Please Please backup your yarn.db
database first! š
Heaven knows I donāt want to be responsible for fucking up a production database here or there š¤£
@prologic@twtxt.net earlier you suggested extending hashes to 11 characters, but hereās an argument that they should be even longer than that.
Imagine I found this twt one day at https://example.com/twtxt.txt :
2024-09-14T22:00Z Useful backup command: rsync -a ā$HOMEā /mnt/backup
and I responded with ā(#5dgoirqemeq) Thanks for the tip!ā. Then Iāve endorsed the twt, but it could latter get changed to
2024-09-14T22:00Z Useful backup command: rm -rf /some_important_directory
which also has an 11-character base32 hash of 5dgoirqemeq. (Iām using the existing hashing method with https://example.com/twtxt.txt as the feed url, but Iām taking 11 characters instead of 7 from the end of the base32 encoding.)
Thatās what I meant by āspoofingā in an earlier twt.
I donāt know if preventing this sort of attack should be a goal, but if it is, the number of bits in the hash should be at least two times log2(number of attempts we want to defend against), where the ātwo timesā is because of the birthday paradox.
Side note: current hashes always end with āaā or āqā, which is a bit wasteful. Maybe we should take the first N characters of the base32 encoding instead of the last N.
Code I used for the above example: https://fossil.falsifian.org/misc/file?name=src/twt_collision/find_collision.c
I only needed to compute 43394987 hashes to find it.
@prologic@twtxt.net How does yarn.socialās API fix the problem of centralization? I still need to know whose API to use.
Say I see a twt beginning (#hash) and I want to look up the start of the thread. Is the idea that if that twt is hosted by a a yarn.social pod, it is likely to know the thread start, so I should query that particular pod for the hash? But what if no yarn.social pods are involved?
The community seems small enough that a registry server should be able to keep up, and I can have a couple of others as backups. Or I could crawl the list of feeds followed by whoever emitted the twt that prompted my query.
I have successfully used registry servers a little bit, e.g. to find a feed that mentioned a tag I was interested in. Was even thinking of making my own, if I get bored of my too many other projects :-)
@mckinley@twtxt.net for me:
- a wall mount 6U rack which has:
- 1U patch panel
- 1U switch
- 2U UPS
- 1U server, intel atom 4G ram, debian (used to be main. now just has prometheus)
- 1U patch panel
- a mini ryzon 16 core 64G ram, fedora (new main)
- multiple docker services hosted.
- multiple docker services hosted.
- synology nas with 4 2TB drives
- turris omnia WRT router -> fiber uplink
network is a mix of wireguard, zerotier.
- wireguard to my external vms hosted in various global regions.
- this allows me ingress since my ISP has me behind CG-NAT
- this allows me ingress since my ISP has me behind CG-NAT
- zerotier is more for devices for transparent vpn into my network
i use ssh and remote desktop to get in and about. typically via zerotier vpn. I have one of my VMs with ssh on a backup port for break glass to get back into the network if needed.
everything has ipv6 though my ISP does not provide it. I have to tunnel it in from my VMs.
Iāve made smart and added a backup copy of my twtxt file to Storj cos my internet keeps on dying
An official FBI document dated January 2021, obtained by the American association āProperty of Peopleā through the Freedom of Information Act.
This document summarizes the possibilities for legal access to data from nine instant messaging services: iMessage, Line, Signal, Telegram, Threema, Viber, WeChat, WhatsApp and Wickr. For each software, different judicial methods are explored, such as subpoena, search warrant, active collection of communications metadata (āPen Registerā) or connection data retention law (ā18 USC§2703ā). Here, in essence, is the information the FBI says it can retrieve:
Apple iMessage: basic subscriber data; in the case of an iPhone user, investigators may be able to get their hands on message content if the user uses iCloud to synchronize iMessage messages or to back up data on their phone.
Line: account data (image, username, e-mail address, phone number, Line ID, creation date, usage data, etc.); if the user has not activated end-to-end encryption, investigators can retrieve the texts of exchanges over a seven-day period, but not other data (audio, video, images, location).
Signal: date and time of account creation and date of last connection.
Telegram: IP address and phone number for investigations into confirmed terrorists, otherwise nothing.
Threema: cryptographic fingerprint of phone number and e-mail address, push service tokens if used, public key, account creation date, last connection date.
Viber: account data and IP address used to create the account; investigators can also access message history (date, time, source, destination).
WeChat: basic data such as name, phone number, e-mail and IP address, but only for non-Chinese users.
WhatsApp: the targeted personās basic data, address book and contacts who have the targeted person in their address book; it is possible to collect message metadata in real time (āPen Registerā); message content can be retrieved via iCloud backups.
Wickr: Date and time of account creation, types of terminal on which the application is installed, date of last connection, number of messages exchanged, external identifiers associated with the account (e-mail addresses, telephone numbers), avatar image, data linked to adding or deleting.
TL;DR Signal is the messaging system that provides the least information to investigators.
Follow-up question for you guys: Where do you backup your files to? Anything besides the local NAS?
@mckinley@twtxt.net ninja backup and Borg
@mckinley@twtxt.net Yeah, thatās more clear. š
Systems that are on all the time donāt benefit as much from at-rest encryption, anyway.
Right, especially not if itās ācloud storageā. š (Weāre only doing it on our backup servers, which are ārealā hardware.)
restic Ā· Backups done right! ā In case no-one has used this wonderful tool restic
yet, I can beyond a doubt assure you it is really quite fantastic š #backups
The battery life in this i9 MacBook Pro from 2018 has diminished to being barely enough to serve as an UPS backup system with enough time to perform a safe shutdown
Getting started with Restic for my backup needs. Itās really neat and fast.
Lost my recent links collection from links.oevl.info. MAKE FUCKING BACKUPS #ShitHappens
Saved Game History: Battery Backups, Memory Cards, and the Cloud https://tedium.co/2019/02/21/video-game-save-state-history/
How to Backup Your Allo Conversations Before Google Shuts it DownĀ https://lifehacker.com/how-to-backup-your-allo-conversations-before-google-shu-1830922541
Somehow, http://www.lord-enki.net/medium-backup/2016-09-01_A-Qualified-Defense-of-JargonāOther-In-Group-Signifiers-2fe2cd37b66b.html is getting a lot of hate. I thought it was pretty even-handed. Do people dislike it because I said culture fit should only matter when it impacts effectiveness, or that itās useful at all?
In case anybody cares, Iāve fixed up my mirror of all my medium posts. (Useful if you are a cheapskate!) As always, I prefer people with a medium account to actually go there & clap so I get a dime. http://www.lord-enki.net/medium-backup/
Against trendism: ipfs://QmQDqrz8Asn3wPbiTHFH9pyAXPNxwbeytJgzWUHF1PZup2 / http://ipfs.io/ipfs/QmQDqrz8Asn3wPbiTHFH9pyAXPNxwbeytJgzWUHF1PZup2 / gopher://fuckup.solutions/1enkiv2/medium-backup/2018-04-01_Against-trendismāhow-to-defang-the-social-media-disinformation-complex-81a8e2635956.txt / https://medium.com/@/against-trendism-how-to-defang-the-social-media-disinformation-complex-81a8e2635956
Kaukatcr: an experiment in language design for multi-dimensional spaces http://www.lord-enki.net/medium-backup/2018-04-12_Kaukatcrāan-experiment-in-language-design-for-multi-dimensional-spaces-cc038caafff9.html
Building a backup bee - Food and Environment Reporting Network https://thefern.org/2018/02/building-backup-bee/
The thing about HTTP is that you should only get a 404 if two meteors simultaneously wiped out both your primary and backup copy (or if it never existed in the first place)
tumblr-utils/tumblr_backup.md at master Ā· bbolli/tumblr-utils Ā· GitHub https://github.com/bbolli/tumblr-utils/blob/master/tumblr_backup.md