• 1 Post
  • 164 Comments
Joined 1 year ago
cake
Cake day: June 10th, 2023

help-circle




  • I’m not a lawyer, but since Lemmy instances aren’t “professional or commercial activity”, I doubt a GPDR request would be applicable.

    Some people who run instances might have the ability to do some sort of database export for a specific user, but the vast majority of us are just barely technical enough to keep Lemmy running and updated.

    The last time I touched our database I accidentally wiped out all data older than 1 month and had to restore a backup.

    I think we do have the option to remove a user by purging them through the UI, but an export isn’t an option at this point.











  • I use zfs with Proxmox. I have it as a bind mount to Turnkey Fileserver (a default lxc template).

    I access everything through NFS (via turnkey Fileserver). Even other VMs just get the NFS added to the fstab file. File transfers happen extremely fast VM to VM, even though it’s “network” storage.

    This gives me the benefits of zfs, and NFS handles the “what if’s”, like what if two VMs access the same file at the same time. I don’t know exactly what NFS does in that case, but I haven’t run into any problems in the past 5+ years.

    Another thing that comes to mind is you should make turnkey Fileserver a privileged container, so that file ownership is done through the default user (1000 if I remember correctly). Unprivileged uses wonky UIDs which requires some magic config which you can find in the docs. It works either way, but I chose the privileged route. Others will have different opinions.





  • There are two types, CMR and SMR. You can read online about the differences. CMR is better because SMR tries to be all fancy in order to increase capacity, but at the cost of speed and data integrity.

    It won’t be front and center in the specs of a particular drive, but you usually find the info somewhere.

    I wouldn’t worry about higher capacity failing sooner. If you have 10x4TB vs 2x20TB, that’s 5x as many drives to go bad. So a 20TB drive would need a 5x worse fail rate to be considered worse. A pro of larger (fewer) drives is lower power consumption. 5-10 watts per drive doesn’t sound like much, but it adds up.