• 0 Posts
  • 60 Comments
Joined 2 years ago
cake
Cake day: April 23rd, 2023

help-circle

  • I would recommend against pairing Battlemage with a low-spec CPU. As shown by Hardware Canucks, Hardware Unboxed, and others, Intel’s Arc graphics driver overhead is currently much higher than competitors, which means they’re disproportionately affected by having a weaker CPU. This causes the B580 to lose significantly more performance when paired with low-end CPUs than a roughly equivalent Nvidia or AMD card. At the very low end, the difference is especially stark. In some games, the B580 goes from neck-and-neck with a 4060 on a high-end CPU to losing half its performance with a low-end older CPU, while the 4060 only loses about 25%.

    If you’re really stuck with a lower-end CPU, it would be far better to get a used midrange AMD or Nvidia GPU from an older product generation for the same price and use that.



  • Onihikage@beehaw.orgtoLinux@lemmy.mlLinux suggestion
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    2 months ago

    Have you ever seen Linux Journey? It’s a very informative set of tutorials on how Linux fundamentally works under the hood; all the separate systems that together create an operating system. The concepts you learn there will apply to almost any distro in some way, even if some distros (like Atomic ones) don’t let you mess with all of it.

    For more top-level transition concerns, given that you’re coming from stock Debian running KDE… Bazzite can also run KDE, so provided you select KDE when you download it, your GUI experience should be pretty much identical. Some minor but important differences would include themes, but there are guides for that, too.

    When it comes to package management, the intent on Atomic systems is you basically don’t install traditional packages (Flatpaks are the preferred option), but Bazzite has frameworks in place such that you can install pretty much any package from any distro, as laid out in their documentation I linked in my previous post and just now. Work is also ongoing to make traditional package-based software installations more seamless with an incoming switch from rpm-ostree to bootc, but that’s getting into the weeds. If you have a deb file for a GUI program that’s not available as a Flatpak, you’ll be using a Distrobox to install it.

    If you have any specific concerns about the differences, let me know and I can hopefully give you more details.


  • Onihikage@beehaw.orgtoLinux@lemmy.mlLinux suggestion
    link
    fedilink
    English
    arrow-up
    2
    ·
    3 months ago

    I can highly recommend Bazzite for your needs. It has a KDE version which is clearly your favorite Desktop Environment (DE), it’s extremely safe/stable due to being an Atomic distro (you can always boot into the previous image if a system update broke something), has incredible documentation, supports almost any traditional app through Distrobox (VPN requires rpm-ostree for now), has a scripted easy install of Waydroid for native android emulation, and has a few tweaks preconfigured to ensure the desktop gaming experience is a little more seamless out of the box than a stock distro. It really seems to tick all the boxes for what you’re looking for.

    If you want more focus on development and less on gaming, the Universal Blue team also makes Aurora for more developer-focused workloads, but Steam not being included in the image does introduce some usability regressions - Steam running via Flatpak or Distrobox is just plain less capable than a native install, though work is ongoing to make native installs Just Work even on Atomic systems.




  • In general, Bazzite being immutable just means the core system isn’t modular to the end user to the degree that Arch is. You of course can use flatpaks or appimages like any distro, and there are still several ways to install traditional rpm/deb/aur programs (the usual Fedora method doesn’t work because dnf doesn’t exist). If it’s just an app that doesn’t require significant integration with the OS, the recommendation is to install them into a distrobox container (where dnf does exist) and then distrobox-export [program] to make them visible to the host system. VPNs need a little more integration so those are installed by layering with rpm-ostree and then enabling the systemd service(s). Layering makes updates take longer to install so it should be avoided when possible.

    One of the interesting things about Universal Blue’s images like Bazzite is if you want the benefits of atomic while also having a more custom system than they offer without having to install a bunch of things in rpm-ostree, the process to build a custom image based on one of theirs is apparently quite easy to do and automate, though I haven’t done it myself.


  • In general, yes. Most of the difficulty is due to being on Linux and running games through the Proton/WINE compatibility layer, so there can be an extra layer of jank involved, but it’s very possible.

    If modding consists of dropping files into the game directory, it will work almost exactly the same as in Windows. However, if some of those files replace the game’s DLLs, then whatever WINE runner you use might need to be told to use the DLLs in the game directory instead of its own.

    If you need to use a mod manager, that situation is still not ideal - native Linux mod managers I know of are only the Nexus Mods app (very new, there’s some talk of it being integrated directly into the Heroic launcher) and Limo. Everything else, you’ll be running whatever bespoke Windows mod manager your game uses through Proton/WINE, probably with Steam Tinker Launch, possibly Lutris.

    tl;dr There can be an extra layer of complexity over modding on Windows, but it’s otherwise comparable.


  • During boot, you’re presented with 4 snapshots you can choose between so if an update did happen to break something, it’s easy as just choosing an older snapshot after a reboot.

    Those are actually just two snapshots, there’s a bug in GRUB that displays them twice. Purely visual, and you can fix it with a ujust script, run in the terminal with ujust configure-grub. There are lots of little scripted tweaks and installations available; you can get most of the list by running ujust by itself. Incredible work by the maintainers.




  • You’re entirely correct, but in theory they can give it a pretty good go, it just requires a lot more computation, developer time, and non-LLM data structures than these companies are willing to spend money on. For any single query, they’d have to get dozens if not hundreds of separate responses from additional LLM instances spun up on the side, many of which would be customized for specific subjects, as well as specialty engines such as Wolfram Alpha for anything directly requiring math.

    LLMs in such a system would be used only as modules in a handcrafted algorithm, modules which do exactly what they’re good at in a way that is useful. To give an example, if you pass a specific context to an LLM with the right format of instructions, and then ask it a yes-or-no question, even very small and lightweight models often give the same answer a human would. Like this, human-readable text can be converted into binary switches for an algorithmic state machine with thousands of branches of pre-written logic.

    Not only would this probably use an even more insane amount of electricity than the current approach of “build a huge LLM and let it handle everything directly”, it would take much longer to generate responses to novel queries.




  • OP is absolutely mistaken that it’s somehow ableist to stick to a meeting deadline or similar “punishment” for lateness, and t3rmit3 has said why much more eloquently than I could. However, you’ve said something that I can’t let pass without a rebuttal.

    perpetual lateness means someone values their time more than they do the commitment and the time of others. period.
    […]
    perpetual lateness, though, is a statement, that individual could not give a shit what others needs and responsibilities are

    This is making a moral judgment on what you believe is in someone’s mind, and your judgment is based on a false premise. There exists an extremely common mental disorder (so common that some might consider it a form of neurodivergence) that when left untreated makes it much harder to do the things you want and are obligated to do. It’s harder to start doing things, it’s harder to stop, it’s harder to focus yet too easy to focus, it’s harder to remember important things, and it’s harder to motivate yourself to do anything you aren’t doing at any given moment, and anything you have to put effort into motivating yourself to do leaves you with less mental energy to do anything else in that category.

    The one thing that can usually overcome all of these mental blocks is panic - when you’re actually out of time and Consequences are approaching if you don’t do something RIGHT NOW then you can finally do what you need to do and get something done - later than you wanted, worse than you wanted, more mentally drained, and with plenty of reasons to beat yourself up over it, not that it helps if you do. This is the reason behind why most people show up perpetually late. They might not let the emotional turmoil show, but if they’re consistently a few minutes late for everything, I can just about promise it’s not because they don’t care.

    People who have this disorder and receive prescription medication for it often describe the first dose as like receiving superpowers. The idea that they can decide they want to do something, and then just go do it? Without thinking about it? No buildup? No psyching yourself into it? No roundabout coping strategies? No reorganizing the entire structure of your life to make it happen? No bargaining with the goddamn monkey in your brain that almost never lets you do the rational thing? Wait, normal people don’t have the monkey? They live like this every day, without any expensive pills? Impossible. It couldn’t be that simple. Do they have any idea how lucky they are?

    Your misplaced sense of moral superiority is unfortunately quite common, but it’s not going to help these people, it’s going to hurt them. If it’s affecting their life, and it often is, they need treatment and training in how their brain works, not to be told they’re a piece of shit who doesn’t care about others and are choosing to inconvenience everyone else in their life including themselves. That’s only going to put them in a worse place.


  • Unfortunately I can’t even test Llama 3.1 in Alpaca because it refuses to download, showing some error message with the important bits cut off.

    That said, the Alpaca download interface seems much more robust, allowing me to select a model and then select any version of it for download, not just apparently picking whatever version it thinks I should use. That’s an improvement for sure. On GPT4All I basically have to download the model manually if I want one that’s not the default, and when I do that there’s a decent chance it doesn’t run on GPU.

    However, GPT4All allows me to plainly see how I can edit the system prompt and many other parameters the model is run with, and even configure multiple sets of parameters for the same model. That allows me to effectively pre-configure a model in much more creative ways, such as programming it to be a specific character with a specific background and mindset. I can get the Mistral model from earlier to act like anything from a very curt and emotionally neutral virtual intelligence named Jarvis to a grumpy fantasy monster whose behavior is transcribed by a narrator. GPT4All can even present an API endpoint to localhost for other programs to use.

    Alpaca seems to have some degree of model customization, but I can’t tell how well it compares, probably because I’m not familiar with using ollama and I don’t feel like tinkering with it since it doesn’t want to use my GPU. The one thing I can see that’s better in it is the use of multiple models at the same time; right now GPT4All will unload one model before it loads another.


  • I have a fairly substantial 16gb AMD GPU, and when I load in Llama 3.1 8B Instruct 128k (Q4_0), it gives me about 12 tokens per second. That’s reasonably fast enough for me, but only 50% faster than CPU (which I test by loading mlabonne’s abliterated Q4_K_M version, which runs on CPU in GPT4All, though I have no idea if that’s actually meant to be comparable in performance).

    Then I load in Nous Hermes 2 Mistral 7B DPO (also Q4_0) and it blazes through at 50+ tokens per second. So I don’t really know what’s going on there. Seems like performance varies a lot from model to model, but I don’t know enough to speculate why. I can’t even try Gemma2 models, GPT4All just crashes with them. I should probably test Alpaca to see if these perform any different there…



  • PCIe gen 5 is for the PCIe slots and NVMe storage slots, but they’re backwards compatible; you can put a gen 3 component in a gen 5 slot and it will work at gen 3 speeds. Similarly, if you put a gen 5 component in a gen 4 slot, it will be limited to gen 4 speeds. Right now there’s very little appreciable difference between gen 4 and gen 5 unless you’re spending a lot of money on the component (GPU/storage). Another thing to note is that Gen 5 requires that both the CPU and motherboard support it; a CPU with gen 4 support in a gen 5 motherboard will limit all the slots to gen 4 speeds.

    RAM is a totally different standard that must be matched exactly for what the motherboard has; if it’s a DDR5 motherboard then you have to use DDR5 RAM or it won’t even fit in the slots. You can get a PCIe gen 5 motherboard and just use gen 4 SSDs or GPUs, that’s perfectly fine and leaves you room to upgrade later.