• 1 Post
  • 140 Comments
Joined 2 years ago
cake
Cake day: March 22nd, 2024

help-circle
  • Well Mint is technically fine, right? Their Nvidia support is 1st party, so it should work out of the box.

    Pretty sure Ubuntu does too.

    Debian, specifically, does not though. And I’m not sure how ‘behind’ Mint and Ubuntu are on their DE and Nvidia driver packages these days, which could be an issue sometimes. But I think many remember Ubuntu/Mint from older days when they were worse in this regard.


  • Nobarra, Bazzite, or CachyOS.

    I’d say Nobara or Bazzite are better for ‘I install it and it just works.’

    Cachy is better for the learning aspect. It’s not hard, but there are more choices to make, and you’re closer to the Arch wiki and all its excellent resources/tutorials.

    I am biased, as I run CachyOS and I love it. I also love how much stuff is in its repos, including everything you need to game optimally, and how easy CUDA is (which is part of what you need for CAD).


    Whatever you choose, do not, I repeat DO NOT install Fedora, Debian, or anything that doesn’t explicitly support Nvidia laptops by default, out of the box, or you are in for a world of pain. If any guide starts with ‘install these 3rd party repos’ or so, you have entered a danger zone, and you will hate linux.


  • brucethemoose@lemmy.worldtoLinux@lemmy.mlNeed text editor advice
    link
    fedilink
    arrow-up
    14
    arrow-down
    1
    ·
    edit-2
    1 month ago

    VSCodium, or some similar VSCode build/derivative.

    I know, I know, but the critical mass is just so useful. As a random example, there are specific extensions to support game modding in Paradox scripting language, or Rimworld XML. Nothing else has so many random niches filled.

    It’s fast with big files (faster than anything I’ve tried other than ‘specialized’ log readers and such), it’s a fast search, it’s got good git support, it’s got support for sudo file editing…


  • This is interesting because theres not a ton of direct Windows vs. linux game benchmarking, and now there’s about to be. GN churns though a lot of hardware and testing.

    And excellent, because being linux, drawing attention to issues increases the chances of them getting fixed, whereas that is hardly the case for Windows.

    Arch (with KDE I presume?) + Bazzite is not bad either. There’s a lot of handwaving over they should have chosen this or that distro, but they’re both very popular in the gaming space, so I feel that’s fairly representative of many distros.







  • Yeah. But it also messes stuff up from the llama.cpp baseline, and hides or doesn’t support some features/optimizations, and definitely doesn’t support the more efficient iq_k quants of ik_llama.cpp and its specialzied MoE offloading.

    And that’s not even getting into the various controversies around ollama (like broken GGUFs or indications they’re going closed source in some form).

    …It just depends on how much performance you want to squeeze out, and how much time you want to spend on the endeavor. Small LLMs are kinda marginal though, so IMO its important if you really want to try; otherwise one is probably better off spending a few bucks on an API that doesn’t log requests.




  • At risk of getting more technical, ik_llama.cpp has a good built in webui:

    https://github.com/ikawrakow/ik_llama.cpp/

    Getting more technical, its also way better than ollama. You can run models way smarter than ollama can on the same hardware.

    For reference, I’m running GLM-4 (667 GB of raw weights) on a single RTX 3090/Ryzen gaming rig, at reading speed, with pretty low quantization distortion.

    And if you want a ‘look this up on the internet for me’ assistant (which you need for them to be truly useful), you need another docker project as well.

    …That’s just how LLM self hosting is now. It’s simply too hardware intense and ad hoc to be easy and smart and cheap. You can indeed host a small ‘default’ LLM without much tinkering, but its going to be pretty dumb, and pretty slow on ollama defaults.



  • Oh wow, that’s awesome! I didn’t know folks ran TDP tests like this, just that my old 3090 seems to have a minimum sweet spot around that same same ~200W based on my own testing, but I figured the 4000 or 5000 series might go lower. Apparently not, at least for the big die.

    I also figured the 395 would draw more than 55W! That’s also awesome! I suspect newer, smaller GPUs like the 9000 or 5000 series still make the value proposition questionable, but still you make an excellent point.

    And for reference, I just checked, and my dGPU hovers around 30W idle with no display connected.


  • brucethemoose@lemmy.worldtoSelfhosted@lemmy.world1U mini PC for AI?
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    2 months ago

    Eh, but you’d be way better off with an X3D CPU in that scenario, which is both significantly faster in games, about as fast outside them (unless you’re dram bandwidth limited) and more power efficient (because they clock relatively low).

    You’re right about the 395 being a fine HTPC machine by itself.

    But I’m also saying even an older 7900, 4090 or whatever would be way lower power at the same performance as the 395’s IGP, and whisper quiet in comparison. Even if cost is no object. And if that’s the case, why keep a big IGP at all? It just doesn’t make sense to pair them without some weirdly specific use case that can use both at once, or that a discrete GPU literally can’t do because it doesn’t have enough VRAM like the 395 does.


  • brucethemoose@lemmy.worldtoSelfhosted@lemmy.world1U mini PC for AI?
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    2 months ago

    Eh, actually that’s not what I had in mind:

    • Discrete desktop graphics idle hot. I think my 3090 uses at least 40W doing literally nothing.

    • It’s always better to run big dies slower than small dies at high clockspeeds. In other words, if you underclocked a big desktop GPU to 1/2 its peak clockspeed, it would use less than a fourth of the energy and run basically inaudible… and still be faster than the iGPU. So why keep a big iGPU around?

    My use case was multitasking and compute stuff. EG game/use the discrete GPU while your IGP churns away running something. Or combine them in some workloads.

    Even the 395 by itself doesn’t make a ton of sense for an HTPC because AMD slaps so much CPU on it. It’s way too expensive and makes it power thirsty. A single CCD (8 cores instead of 16) + the full integrated GPU would be perfect and lower power, but AMD inexplicably does not offer that.

    Also, I’ll add that my 3090 is basically inaudible next to a TV… key is to cap its clocks, and the fans barely even spin up.



  • brucethemoose@lemmy.worldtoSelfhosted@lemmy.world1U mini PC for AI?
    link
    fedilink
    English
    arrow-up
    4
    ·
    edit-2
    2 months ago

    It’s PCIe 4.0 :(

    but these laptop chips are pretty constrained lanes wise

    Indeed. I read Strix Halo only has 16 4.0 PCIe lanes in addition to its USB4, which is resonable given this isn’t supposed to be paired with discrete graphics. But I’d happily trade an NVMe slot (still leaving one) for x8.

    One of the links to a CCD could theoretically be wired to a GPU, right? Kinda like how EPYC can switch its IO between infinity fabric for 2P servers, and extra PCIe in 1P configurations. But I doubt we’ll ever see such a product.