Father, Hacker (Information Security Professional), Open Source Software Developer, Inventor, and 3D printing enthusiast

  • 3 Posts
  • 18 Comments
Joined 1 year ago
cake
Cake day: June 23rd, 2023

help-circle


  • You had corruption with btrfs? Was this with a spinning disk or an SSD?

    I’ve been using btrfs for over a decade on several filesystems/machines and I’ve had my share of problems (mostly due to ignorance) but I’ve never encountered corruption. Mostly I just run out of disk space because I forgot to balance or the disk itself had an issue and I lost whatever it was that was stored in those blocks.

    I’ve had to repair a btrfs partition before due to who-knows-what back when it was new but it’s been over a decade since I’ve had an issue like that. I remember btrfs check --repair being totally useless back then haha. My memory on that event is fuzzy but I think I fixed whatever it was bitching about by remounting the filesystem with an extra option that forced it to recreate a cache of some sort. It ran for many years after that until the disk spun itself into oblivion.


  • I wouldn’t say, “repairing XFS is much easier.” Yeah, fsck -y with XFS is really all you have to do 99% of the time but also you’re much more likely to get corrupted stuff when you’re in that situation compared to say, btrfs which supports snapshotting and redundancy.

    Another problem with XFS is its lack of flexibility. By that I don’t mean, “you can configure it across any number of partitions on-the-fly in any number of (extreme) ways” (like you can with btrfs and zfs). I mean it doesn’t have very many options as to how it should deal with things like inodes (e.g. tail allocation). You can increase the total amount of space allowed for inode allocation but only when you create the filesystem and even then it has a (kind of absurdly) limited number that would surprise most folks here.

    As an example, with an XFS filesystem, in order to store 2 billion symlimks (each one takes an inode) you would need 1TiB of storage just for the inodes. Contrast that with something like btrfs with max_inline set to 2048 (the default) and 2 billion symlimks will take up a little less than 1GB (assuming a simplistic setup on at least a 50GB single partition).

    Learn more about btrfs inlining: https://btrfs.readthedocs.io/en/latest/Inline-files.html


  • One point: ext4 has a maximum file size of 16TiB. To a regular user that is stupidly huge and of no concern but it’s exactly the type of thing you overlook if you “just use ext4” on anything and everything then end up with your database broken at work because of said bad advice.

    Use the filesystem that makes the most sense for your use case. Consider it every single time you format a disk. Don’t become complacent! Also fuck around with the new shit from time to time! I decided to format my Linux desktop partitions with btrfs over a decade ago and as a result I’m an excellent user of that filesystem but you know what? I’m thinking I’ll try bcachefs soon and fiddle around more with my zfs partition on my HTPC.

    BTW: If you’re thinking about trying out btrfs I would encourage you to learn about it’s non-trivial maintenance tasks. btrfs needs you to fuck with it from time to time or you’ll run out of disk space “for no reason”. You can schedule cron jobs to take care of everything (as I have done) but you still need to learn how it all works. It’s not a “set it and forget it” FS like ext4.





  • There’s not much to learn in Windows land! Learn how to set file permissions, how the registry works (and some important settings that use it), and how Active Directory works (it’s LDAP) and you’ll be fine.

    If you’re used to using Linux nothing will frustrate you more than being forced to use a Windows desktop. The stuff you use every day just isn’t there. You can add on lots of 3rd party tools to make it better but it’ll never measure up.

    When you have to go out on the Internet to download endless amounts of 3rd party tools the security alarms in your head might start going off. Windows users have just learned over time to ignore them 🤣


  • I interview developers and information security people all the time. I always ask lots of questions about Linux. As far as I’m concerned:

    • If you’re claiming to be an infosec professional and don’t know Linux you’re a fraud.
    • If you’re a developer and you don’t know how to deploy to Linux servers you’re useless.

    So yeah: Get good with Linux. Especially permissions! Holy shit the amount of people I interview that don’t know basic Linux permissions (or even about file permissions in general) is unreal.

    Like, dude: Have you just been chmod 777 everything all this time? WTF! Immediate red flag this guy cannot be trusted with anything.



  • You say that because you don’t realize the benefits:

    • Better support for Linux with any new PC hardware on day 1. This includes things like USB devices, monitors, KVMs, UPS, everything.
    • Better support for all commercial software in general. More software will become available and it’ll be higher quality.
    • Vendors will be forced to test all their stuff on Linux which means it’ll all become more reliable and less glitchy.
    • There will be more diversity in software and distros which means widespread attacks (aka hacking, worms, viruses, etc) will have less success and smaller impacts.
    • The more Linux users there are the more Linux developers will result. It’s also much easier to start learning how to code on a Linux desktop than it is in Windows.
    • Better security for the entire world. Linux has a vastly superior security architecture than Windows and a vastly superior track record. The more Linux users there are, the harder it will be for malicious entities to break into their PCs which translates into a more secure world.
    • It’s much easier (for experienced users) to troubleshoot and fix problems in Linux than in Windows. This will lead to support teams everywhere getting frustrated whenever they have to deal with Windows users (this is already the case for many software vendors, haha). Therefore, it makes support people happy and easy going. Who doesn’t want to reach a happy, helpful person for technical support instead of the usual defiant/adversarial support tech? 😁
    • The worst sorts of hardware vendors won’t be able to get away with their usual bullshit. For example, if there were enough Linux users HP wouldn’t be offering extremely invasive 2GB printer “drivers” because their Windows customers would know enough Linux users that they’d be rightfully pissed and not depressively submissive like they are now.
    • When you do have a problem it will be easier to find a solution because the likelihood that someone else already had it and posted a solution will be higher (though admittedly this factor doesn’t seem to do much for Windows currently because of how obtuse and obfuscated everything is in that OS).

    There’s actually a lot more reasons but that’s probably enough for now 😁



  • I’d love to see more adoption of… I2C!

    Bazillions of motherboards and SBCs support I2C and many have the ability to use it via GPIO pins or even have connectors just for I2C devices (e.g. QWIIC). Yet there’s very little in the way of things you can buy and plug in. It feels like such a waste!

    There’s all sorts of neat and useful things we could plug in and make use of if only there were software to use it. For example, cheap color sensors, nifty gesture sensors, time-of-flight sensors, light sensors, and more.

    There’s lmsensors which knows I2C and can magically understand zillions of temperature sensors and PWM things (e.g. fan control). We need something like that for all those cool devices and chips that speak I2C.



  • This is caused by your root controller’s limited bandwidth and it’s inability to handle that many 3.0 devices at the same time. Some of the newer motherboards with USB C PD have controllers in them that can do a lot more.

    It’s basically a hack on part of the company that made the root controller IC. They know they only have enough internal bandwidth to support 16 USB 3.0 devices so they intentionally bork things when you plug in more than that since their Transaction Translator (TT) can’t handle more and they were too lazy to bother implementing the ability to share 2.0 and 3.0 properly.

    I’m guessing the decision went something like this…

    “We have enough bandwidth for 16 3.0 devices… What do we do if someone plugs in more than that?” “Only a few people will ever have that many! We don’t have the budget to handle every tiny little use case! Just ship it.”

    So it’s not Linux fault in this case. Or at least, if it is (a problem with the driver) it’s because of some proprietary bullshit that the driver requires to function properly 🤷