It’s called Lemmy-Safety of Fedi-Safety depending on where you look.
One thing to note, I wasn’t able to get it running on a VPS because it requires some sort of GPU.
It’s called Lemmy-Safety of Fedi-Safety depending on where you look.
One thing to note, I wasn’t able to get it running on a VPS because it requires some sort of GPU.
Yeah, it’s just something like “Tell us why you want to join this instance”. If the answer is “to promote my content” or “qq”, for example, they don’t get approved.
It’s done by the Lemmy software.
We require applications, and most applications we get are extremely low effort and we don’t approve them. If you have open registrations you’ll be doing a lot of moderation for spam.
Run the software that scans images for CSAM. It’s not perfect but it’s something. If your instance freely hosts whatever without any oversight, word will spread and all of a sudden you’re hosting all sorts of bad stuff. It’s not technically illegal if you don’t know about it, but I personally don’t want anything to do with that.
I’m not a lawyer, but since Lemmy instances aren’t “professional or commercial activity”, I doubt a GPDR request would be applicable.
Some people who run instances might have the ability to do some sort of database export for a specific user, but the vast majority of us are just barely technical enough to keep Lemmy running and updated.
The last time I touched our database I accidentally wiped out all data older than 1 month and had to restore a backup.
I think we do have the option to remove a user by purging them through the UI, but an export isn’t an option at this point.
Looking forward to trying the latest update.
There’s a difference in how active users are counted now, too, so that skews the numbers. An active user used to be a user who posted or commented. Now voting makes a user active.
It sure is pretty.
This looks like a New England forest.
Probably one of those asshole children who’ve only been cycling for a year or two. They don’t take the sport seriously.
And as a bonus, presumably you have a nice file filled with historic dates and times!
I was lucky, too. I’ve waited so long to see these and it never worked out until now.
Block lemmy.ml and hexbear to start. I find their users to be quite toxic.
I switched from portainer to dockge. Dockge makes updating a 1-click process which I love. Portainer is overkill for homelab, but I like how it lists things like images and networks.
I use zfs with Proxmox. I have it as a bind mount to Turnkey Fileserver (a default lxc template).
I access everything through NFS (via turnkey Fileserver). Even other VMs just get the NFS added to the fstab file. File transfers happen extremely fast VM to VM, even though it’s “network” storage.
This gives me the benefits of zfs, and NFS handles the “what if’s”, like what if two VMs access the same file at the same time. I don’t know exactly what NFS does in that case, but I haven’t run into any problems in the past 5+ years.
Another thing that comes to mind is you should make turnkey Fileserver a privileged container, so that file ownership is done through the default user (1000 if I remember correctly). Unprivileged uses wonky UIDs which requires some magic config which you can find in the docs. It works either way, but I chose the privileged route. Others will have different opinions.
Thanks for the suggestion. I ended up using a Raspberry Pi and an old computer monitor to run MagigMirror and MMM-ImmichSlideShow.
I tried ImmichFrame, too, and will revisit it in the future. For now MMM-ImmichSlideShow is working well.
The developer is still active with their other main project, Uptime Kuma. So that’s good.
That’s good
There are two types, CMR and SMR. You can read online about the differences. CMR is better because SMR tries to be all fancy in order to increase capacity, but at the cost of speed and data integrity.
It won’t be front and center in the specs of a particular drive, but you usually find the info somewhere.
I wouldn’t worry about higher capacity failing sooner. If you have 10x4TB vs 2x20TB, that’s 5x as many drives to go bad. So a 20TB drive would need a 5x worse fail rate to be considered worse. A pro of larger (fewer) drives is lower power consumption. 5-10 watts per drive doesn’t sound like much, but it adds up.
Good question, and I’m curious what the experts say. Surely it depends on the software that handles DHCP.
I’ve always set static addresses in the DHCP address range and it has always been reserved and never assigned to other devices. I’ve used ASUS and MikroTik for what it’s worth.
If you’re the type to set static addresses on the devices themselves, then that would certainly increase the risk of a conflict if it’s inside the address range.
I just checked and we have that turned on, too.
We don’t get a lot of applications. A couple per week, maybe.