I see the raise of popularity of Linux laptops so the hardware compatibility is ready out of the box. However I wonder how would I build PC right know that has budget - high end specification. For now I’m thinking

  • Case: does not matter
  • Fans: does not matter
  • PSU: does not matter
  • RAM: does not matter I guess?
  • Disks: does not matter I guess?
  • CPU: AMD / Intel - does not matter but I would prefer AMD
  • GPU: AMD / Intel / Nvidia - for gaming and Wayland - AMD, for AI, ML, CUDA and other first supported technologies - Nvidia.

And now the most confusing part for me - motherboard… Is there even some coreboot or libreboot motherboard for PC that supports “high end” hardware?

Let’s just say also a purpose of this Linux PC. Choose any of these

  1. Blender 3D Animation rendering
  2. Gaming
  3. Local LLM running

If you have some good resources on this also let me know.

  • brucethemoose@lemmy.world
    link
    fedilink
    arrow-up
    12
    ·
    edit-2
    12 hours ago

    Basically the only thing that matters for LLM hosting is VRAM capacity. Hence AMD GPUs can be OK for LLM running, especially if a used 3090/P40 isn’t an option for you. It works fine, and the 7900/6700 are like the only sanely priced 24GB/16GB cards out there.

    I have a 3090, and it’s still a giant pain with wayland, so much that I use my AMD IGP for display output and Nvidia still somehow breaks things. Hence I just do all my gaming in Windows TBH.

    CPU doesn’t matter for llm running, cheap out with a 12600K, 5600, 5700x3d or whatever. And the single-ccd x3d chips are still king for gaming AFAIK.

    • GenderNeutralBro@lemmy.sdf.org
      link
      fedilink
      English
      arrow-up
      3
      ·
      12 hours ago

      Basically the only thing that matters for LLM hosting is VRAM capacity

      I’ll also add that some frameworks and backends still require CUDA. This is improving but before you go and buy an AMD card, make sure the things you want to run will actually run on it.

      For example, bitsandbytes support for non-CUDA backends is still in alpha stage. https://huggingface.co/docs/bitsandbytes/main/en/installation#multi-backend

      • brucethemoose@lemmy.world
        link
        fedilink
        arrow-up
        3
        ·
        11 hours ago

        For local LLM hosting, basically you want exllama, llama.cpp (and derivatives) and vllm, and rocm support for all of them is just fine. It’s absolutely worth having a 24GB AMD card over a 16GB Nvidia one, if that’s the choice.

        The big sticking point I’m not sure about is flash attention for exllama/vllm, but I believe the triton branch of flash attention works fine with AMD GPUs now.

    • Psyhackological@lemmy.mlOP
      link
      fedilink
      arrow-up
      1
      ·
      12 hours ago

      VRAM and RAM I think. Still AMD seems always slower than Nvidia for some reason for this purpose. Same for Blender benchmarks.

      Ah I use my AMD GPU with Bazzite and it is wonderful.

      CPU does not matter when GPU matters. Otherwise small models will do fine on CPU especially with more recent instructions for running LLMs.