• 2 Posts
  • 322 Comments
Joined 8 months ago
cake
Cake day: June 9th, 2024

help-circle

  • I wouldn’t argue with the dude; he’s got a clear case of bad-faith-itis. What you did was bad, so you shouldn’t have done it, but no I won’t tell you how to fix it.

    The absolute best you could have done is cross-posted to a Mastodon/Bluesky/whatever account as well, but you can’t just always go around yanking the rug out underneath communities especially if you’re in a position where it’s not just lazy shitposting and worthless commentary.

    …that said, you have moved anything you can to being posted somewhere in tandem riiiiiiight?


  • As with all things email, they probably really wanted to make sure that the mails were delivered and thus were using a commercial MTA to ensure that.

    I’d wager, even at 20 or 30 or 40k a year, that’s way less than it’d cost to host infra and have at least two if not three engineers available 24/7 to maintain critical infra.

    Looking at my mail, over the years I’ve gotten a couple hundred email from them around certificates and expirations (and other things), and if you assume there’s a couple million sites using these certs, I could easily see how you’d end up in a situation where this could scale in cost very very slowly, until it’s suddenly a major drain.


  • Very very little. It’s a billion tiny little bits of text, and if you have image caching enabled, then all those thumbnails.

    My personal instance doesn’t cache images since I’m the only one using it (which means a cached image does nobody any good), and i use somewhere less than 20gb a month, though I don’t have entirely specific numbers, just before-lemmy and after-lemmy aggregates.




  • I don’t disagree, but if it’s a case where the janky file problem ONLY appears in Jellyfin but not Plex, then, well, jank or not, that’s still Jellyfin doing something weird.

    No reason why Jellyfin would decide the French audio track should be played every 3rd episode, or that it should just pick a random subtitle track when Plex isn’t doing it on exactly the same files.


  • One thing I ran into, though it was a while ago, was that disk caching being on would trash performance for writes on removable media for me.

    The issue ended up being that the kernel would keep flushing the cache to disk, and while it was doing that none of your transfers are happening. So, it’d end up doubling or more the copy time because the write cache wasn’t actually helping removable drives.

    It might be worth remounting without any caching, if it’s on, and seeing if that fixes the mess.

    But, as I said, this has been a few years, so that may no longer be actively the case.


  • If you share access with your media to anyone you’d consider even remotely non-technical, do not drop Jellyfin in their laps.

    The clients aren’t nearly as good as plex, they’re not as universally supported as plex, and the whole thing just has the needs-another-year-or-two-of-polish vibes.

    And before the pitchfork crowd shows up, I’m using Jellyfin exclusively, but I also don’t have people using it who can’t figure out why half the episodes in a tv season pick a different language, or why the subtitles are somtimes english, and sometimes german, or why some videos occasionally don’t have proper audio (l and r are swapped) and how to take care of all of those things.

    I’d also agree your thought that docker is the right approach to go: you don’t need docker swarm, or kubernetes, or whatever other nonsense for your personal plex install, unless you want to learn those technologies.

    Install a base debian via netinstall, install docker, install plex, done.


  • Timely post.

    I was about to make one because iDrive has decided to double their prices, probably because they could.

    $30/tb/year to $50/tb/year is a pretty big jump, but they were also way under the market price so capitalism gonna capital and they’re “optimizing” or someshit.

    I’ve love to be able to push my stuff to some other provider for closer to that $30, but uh, yeah, no freaking clue who since $60/tb/year seems to be the more average price.

    Alternately, a storage option that’s not S3-based would also probably be acceptable. Backups are ~300gb, give or take, and the stuff that does need S3-style storage I can stuff in Cloudflare’s free tier.







  • The chances of both failing is very rare.

    If they’re sequential off the manufacturing line and there’s a fault, they’re more likely to fail around the same time and in the same manner, since you put the surviving drive under a LOT of stress when you start a rebuild after replacing the dead drive.

    Like, that’s the most likely scenario to lose multiple drives and thus the whole array.

    I’ve seen far too many arrays that were built out of a box of drives lose one or two, and during rebuild lose another few and nuke the whole array, so uh, the thought they probably won’t both fail is maybe true, but I wouldn’t wager my data on that assumption.

    (If you care about your data, backups, test the backups, and then even more backups.)


  • You can find reasonably stable and easy to manage software for everything you listed.

    I know this is horribly unpopular around here, but you should, if you want to go this route, look at Nextcloud. It 's a monolithic mess of PHP, but it’s also stable, tested, used and trusted in production, and doesn’t have a history of lighting user data on fire.

    It also doesn’t really change dramatically, because again, it’s used by actual businesses in actual production, so changes are slow (maybe too slow) and methodical.

    The common complaints around performance and the mobile clients are all valid, but if neither of those really cause you issues then it’s a really easy way to handle cloud document storage, organization, photos, notes, calendars, contacts, etc. It’s essentially (with a little tweaking) the entire gSuite, but self-hosted.

    That said, you still need to babysit it, and babysit your data. Backups are a must, and you’re responsible for doing them and testing them. That last part is actually important: a backup that doesn’t have regular tests to make sure they can be restored from aren’t backups they’re just thoughts and prayers sitting somewhere.




  • It is mostly professional/office use where this make sense. I’ve implemented this (well, a similar thing that does the same thing) for clients that want versioning and compliance.

    I’ve worked with/for a lot of places that keep everything because disks are cheap enough that they’ve decided it’s better to have a copy of every git version than not have one and need it some day.

    Or places that have compliance reasons to have to keep copies of every email, document, spreadsheet, picture and so on. You’ll almost never touch “old” data, but you have to hold on to it for a decade somewhere.

    It’s basically cold storage that can immediately pull the data into a fast cache if/when someone needs the older data, but otherwise it just sits there forever on a slow drive.