Easiest? More like… The only way.
Easiest? More like… The only way.
Is this what burnout looks like?
If you don’t want to be on the bleeding edge and want a distro with longer support, CentOS Stream isn’t bad. Sure, there was some controversy surrounding it, when Red Hat killed the old CentOS. But ignoring that, the distro itself is pretty good and stable.
Much better. SSDs and HDDs do monitor the health of the drives (and you can see many parameters through SMART), while pen drives and SD cards don’t.
Of course, they have their limits which is why raid exists. File systems like ZFS are built on the premise that drives are unreliable. It’s up to you if you want that redundancy. The most important thing to not lose data is to have backups. Ideally at least 3 copies, 1 off site (e.g. on a cloud, or on a disk at some place other than your home).
PhotoRec and TestDisk are probably the best, but they don’t recover file structure.
Fuck up #1: no backups
Fuck up #2: using SD cards for data storage. SD cards and USB drives are ephemeral storage devices, not to be relied on. Most of the time they use file systems like FAT32 which are far less safe than NTFS or ext4. Use reliable storage media, like hard drives.
Fuck up #3: no backups.
The honestly prefer the bottom one than the modern 50 step wizards that take 10 seconds for each page to load, and load an ungodly amount of JS scripts.
A company I worked for was using an ancient bug tracking tool (called Pivotal) that looked like a 90s site. It was so fast and responsive. Later, we moved to something modern. It was 10 times worse, significantly slower and overly complex.
I wouldn’t recommend Optiplexes… HP, Dell, Lenovo pre-builts use proprietary parts making them a pain in the rear to work with. I recommend getting a PC made with standard parts.
Containers are very useful because they isolate the application from the rest of your server.
This solves a lot of problems: no dependency conflicts with your operating system, you can upgrade/downgrade any time you want, no state gets stored on your main system which makes resetting the application when it misbehaves as easy as deleting and recreating the container.
Before containers, changing my host OS (e.g. because ZFS wasn’t properly supported on the distro I was using) meant reinstalling and configuring a lot of shit, which could take days. With docker, I can migrate in 1-2 hours… Just install docker on the new OS, copy over the files, docker compose up a few times and done. The only things left to setup are samba, ssh and a few cron jobs.
My experience wasn’t as bad, but after the third time the database got corrupted during an upgrade I stopped using it.
Self hosting basically means you are running the server application yourself. It doesn’t matter if it’s at home, on a cloud service or anywhere else.
I wouldn’t recommend hosting a social network like lemmy, because you would be legally responsible for all the content served from your servers. That means a lot of moderation work. Also, these types of applications are very demanding in terms of data storage, you end up with an ever growing dataset of posts, pictures etc.
But self hosting is very interesting and empowering. There are a lot of applications you can self host, from media servers (Plex, Jellyfin), personal cloud (like Google Drive) with NextCloud, blocking ads with pihole, sync servers for various apps like Obsidian, password manager BitWarden etc. You can even make your own website by coding it, or using a CMS platform like WordPress.
Check the Awesome Self-hosted list on GitHub, has a ton of great stuff.
And in terms of hardware, any old computer or laptop can be used, just install your favorite server OS (Linux, FreeBSD/OpenBSD, even Windows Server). You can play with virtualization too if you have enough horsepower and memory with ESXI or Proxmox, so you can run multiple severs at once on the same computer.
Also renamed xml, renamed json and renamed sqlite.
Sure, but it’s still really interesting from a historic point of view.