A person with way too many hobbies, but I still continue to learn new things.

  • 5 Posts
  • 233 Comments
Joined 1 year ago
cake
Cake day: June 7th, 2023

help-circle
  • Are you sure about that? Ever hear about this supposed predictable network names in recent linux versions? Yeah those can change too. I was trying to set up a new firewall with two internal NICs plus a 4-port card, and they kept moving around. I finally figured out that if I cold-booted the NICs would come up in one order, and if I warm-booted they would come up in a completely different order (like the ports on the card would reverse which order they were detected). This was completely the fault of systemd because when I installed an older linux and used udev to map the ports, it worked exactly as predicted. These days I trust nothing.


  • Shdwdrgn@mander.xyztoSelfhosted@lemmy.worldHelp with ZFS Array
    link
    fedilink
    English
    arrow-up
    6
    ·
    edit-2
    6 hours ago

    OP – if your array is in good condition (and it looks like it is) you have an option to replace drives one by one, but this will take some time (probably over a period of days). The idea is to remove a disk from the pool by its old name, then re-add the disk under the corrected name, wait for the pool to rebuild, then do the process again with the next drive. Double-check, but I think this is the proper procedure…

    zpool offline poolname /dev/nvme1n1p1

    zpool replace poolname /dev/nvme1n1p1 /dev/disk/by-id/drivename

    Check zpool status to confirm when the drive is done rebuilding under the new name, then move on to the next drive. This is the process I use when replacing a failed drive in a pool, and since that one drive is technically in a failed state right now, this same process should work for you to transfer over to the safe names. Keep in mind that this will probably put a lot of strain on your drives since the contents have to be rebuilt (although there is a small possibility zfs may recognize the drive contents and just start working immediately?), so be prepared in case a drive does actually fail during the process.


  • That is definitely true of zfs as well. In fact I have never seen a guide which suggests anything other than using the names found under /dev/disk/by-id/ or /dev/disk/by-id/uuid and that is to prevent this very problem. If the proper convention is used then you can plug the drives in through any available interface, in any order, and zfs will easily re-assemble the pool at boot.

    So now this begs the question… is proxmox using some insane configuration to create drive clusters using the name they happen to boot up with???







  • Seems like a good opportunity to remind folks about the Kiwix project, which allows you to download local private copies of select information such as Wikipedia. It was originally created to provide offline access to content for countries that were otherwise blocked, but events like this have sparked some recent discussion about archiving older files to preserve history.



  • It’s weird to see T-mobile taking this stance. I switched to them years ago because they were one of the few that supported unlocked phones, and even offered them for sale. Their policies might have changed on this, but I just bought an unlocked phone off Ebay this Summer and all I needed to do was pop my sim card into the new device. Hell I had to specifically install the visual voicemail app because there wasn’t any bloatware on the phone when I got it. So I guess I’m not following what their complaint is about?


  • Most of us can’t afford the sort of disk capacity they use, but it would be really cool if there were a project to give volunteers pieces of the archive so that information was spread out. Then volunteers could specify if they want to contribute a few gigabytes to multiple terabytes of drive space towards the project and the software could send out packets any time the content changes. Hmm this description sounds familiar but I can’t think of what else might be doing something similar – anyone know of anything like that that could be applied to the archive?





  • Did you know that, by default, your email sends information to mailing list platforms about your reading activity? The platform gets to know if you opened the message, and often how far along you’ve read in it.

    What is this shitty email program they’re talking about? Sure, they can embed a 1-pixel tracking image to see when you opened the email (if you allow auto-loading images), but how would they know how much you’ve read unless some incredibly horrible email program actively sends out that data?





  • If you want to do it right, try to get a static IP (you may need to get a business account). If your provider doesn’t provide IPv6 to static IPs, go to some place like Hurricane Electric and get a free IPv6 range pointed to your IPv4 static address.

    Alternatively, you might do a search for any DDNS services that provide IPv6 (I’m not sure if any do?), then that service will fllow your residential address when it changes. Either way I think you’ll have some additional costs you need to weigh against your current hosting provider.


  • Seems like that would be an easy problem to solve… require all edits to have a peer review by someone with a minimum credibility before they go live. I can understand when Wikipedia was new, allowing anyone to post edits or new content helped them get going. But now? Why do they still allow any random person to post edits without a minimal amount of verification? Sure it self-corrects given enough time, but meanwhile what happens to all the people looking for factual information and finding trash?