![](/static/253f0d9b/assets/icons/icon-96x96.png)
![](https://lemmy.world/pictrs/image/8286e071-7449-4413-a084-1eb5242e2cf4.png)
+1 for the main risk to my service reliability being me getting distracted by some other shiny thing and getting behind on maintenance.
+1 for the main risk to my service reliability being me getting distracted by some other shiny thing and getting behind on maintenance.
I love this idea (of just picking something I’m loving each month), it would help me overcome my decision paralysis about who to support.
Yes, a few. Signal (daily use), LetsEncrypt & Certbot (EFF). It’s not enough.
One day I decided I’d spend $x every January (when I do all my other donations) on open source stuff I depend on, and roughly in the proportions I depend on them. It quickly became impossible - I can’t just fund Debian (which I use a lot of in VMs), I’d need to think of all their dependencies, same with NGINX, Node etc etc. The mind boggles.
I need something like a Spotify subscription for open source to assuage my guilt of the great value I extract for my personal use of open source.
Yes, and it’s very phrased based, so you sort of absorb the grammar without a lot of explicit rules.
I started as more “homelab” than “selfhosted” as first - so I was just stuffing around playing with things, but then that seemed sort of pointless and I wanted to run real workloads, then I discovered that was super useful and I loved extracting myself from commercial cloud services (dropbox etc). The point of this story is that I sort of built most of the infrastructure before I was running services that I (or family) depended on - which is where it can become a source of stress rather than fun, which is what I’m guessing you’re finding yourself in.
There’s no real way around this (the pressure you’re feeling), if you are running real services it is going to take some sysadmin work to get to the point where you feel relaxed that you can quickly deal with any problems. There’s lots of good advice elsewhere in this thread about bit and pieces to do this - the exact methods are going to vary according to your needs. Here’s mine (which is not perfect!).
I still have lots of single points of failure - Tailscale, my internet provider, my domain provider etc, but I think I’ve addressed the most common which would be hardware failures at home. My monitoring is also probably sub-par, I’m not really looking at logs unless I’m investigating a problem. Maybe there’s a Netdata or something in my future.
You’ve mentioned that a syncing to a remote server for backups is a step you don’t want to take, if you mean managing your own is a step you don’t want to take, then your solutions are a paid backup service like backblaze or, physically shuffling external USB drives (or extra NASs) back and forth to somewhere - depending on what downtime you can tolerate.
Same with consistent pronunciation in Indonesian - it’s so much better. I feel sorry for little kids learning to read English and getting told to ‘sound it out’. Sure thing, which of the five to nine sounds shall I use for the letter ‘a’?
+1 for Syncthing. I run it on a server at home, then on my MacBook over Tailscale. For web access I run FileBrowser (also over Tailscale) against the same directory.
Lots of good mentions here, but I’m not seeing my fav, Smack the Pony
Wish by Peter Goldsworthy. J.J. has always been more at home in Sign language than in spoken English. Recently divorced, he returns to school to teach Sign. His pupils include the foster parents of a beautiful and highly intelligent ape named Eliza.
Greta Tintin Thunberg
I run two local physical servers, one production and one dev (and a third prod2 kept in case of a prod1 failure), and two remote production/backup servers all running Proxmox, and two VPSs. Most apps are dockerised inside LXC containers (on Proxmox) or just docker on Ubuntu (VPSs). Each of the three locations runs a Synology NAS in addition to the server.
Backups run automatically, and I manually run apt updates on everything each weekend with a single ansible playbook. Every host runs a little golang program that exposes the memory and disk use percent as a JSON endpoint, and I use two instances of Uptime Kuma (one local, and one on fly.io) to monitor all of those with keywords.
So -
The Debian thong made me laugh. Who is buying this? For themselves, their partners? I’m imagining Christmas morning when I’m trying to explain the value of this gift you’ve just opened.
My ‘good reason’ is just that it’s super convenient - for backups and painlessly moving apps around between nodes with all their data.
I would run plain LXCs if people nicely packaged up their web apps as LXC templates and made them available on LXCHub for me to run with lxc compose up
, but they generally don’t.
I guess another alternate future would be if Proxmox added docker container supervision to their web interface, but you’re still not going to have the self-contained neat snapshot system that includes the data.
In theory you should be able to convert an OCI container layer by layer into an LXC, so I bet there’s projects out there that attempt this.
No answer, but just to say I run most of my services with this setup - Docker in a Debian LXC under Proxmox, and don’t have this issue. The containers are ‘privileged’, and I have ‘nesting’ ticked on, but apart from that all defaults.
There are a heap of general “Linux Administration” courses which will patch a lot of holes in the knowledge of almost all self-taught self hosters. I’d been using Linux for a while but didn’t know you could tab to complete file names in commands till I learned it on Udemy ¯_(ツ)_/¯
I routinely run my homelab services as a single Docker inside an LXC - they are quicker, and it makes backups and moving them around trivial. However, while you’re learning, a VM (with something conventional like Debian or Ubuntu) is probably advised - it’s a more common experience so you’ll get more helpful advice when you ask a question like this.
For anyone coming to this later, digests are not presented in a super-understandable way (there’s a good explanation of them in this stack-overflow). There appears to be two ways to get the digest that matches the one on the DockerHub page for the tag:
There’s a good discussion of options for finding out if images have been updated in this r/SelfHosted discussion which is what I was trying to achieve.
I’m on iOS. I’ve been testing a beta of Jello that looks really promising, but as a beta has a bit of distance to go. I’ll check out Feishin though - thanks for the recommendation.
I’d love Jellyfin to turn out to be the solution, but I suspect it’s not, at least yet.
I’ve got three of these little 1L HP’s, one for production, a spare, and one for development. But really, it’s a small load - that list would happily run on an old nuc. The constraint is really memory which I’ve mostly addressed by moving from VMs to LXCs. And I could be even more efficient by just running all the docker containers on one host if I had to.
Storage for media and backups is a Synology NAS.
Yeah na, put your home services in Tailscale, and for your VPS services set up the firewall for HTTP, HTTPS and SSH only, no root login, use keys, and run fail2ban to make hacking your SSH expensive. You’re a much smaller target than you think - really it’s just bots knocking on your door and they don’t have a profit motive for a DDOS.
From your description, I’d have the website on a VPS, and Immich at home behind TailScale. Job’s a goodun.