Fediverse is worse than Reddit. Mod abuse, admin abuse, disinformation, and people simping for literal terrorists.

  • 0 Posts
  • 57 Comments
Joined 11 months ago
cake
Cake day: January 3rd, 2024

help-circle








  • DarkThoughts@fedia.iotolinuxmemes@lemmy.worldLinux is easy!
    link
    fedilink
    arrow-up
    1
    arrow-down
    1
    ·
    29 days ago

    Try to install Mullvad VPN. Don’t use the .rpm because if you happen to be able to install that it bricks your system. In my case that didn’t even work, luckily, but the app just won’t seem to work within a suggested DistroBox setting either. Can’t get ROCm to properly work either, which is already a pain in the ass on a regular distro.



  • ggml_cuda_compute_forward: ADD failed
    CUDA error: shared object initialization failed
      current device: 0, in function ggml_cuda_compute_forward at ggml/src/ggml-cuda.cu:2365
      err
    ggml/src/ggml-cuda.cu:107: CUDA error
    

    I didn’t do anything past using yay to install the AUR koboldcpp-hipblas package, and customtkinter, since the UI wouldn’t work otherwise. The koboldcpp-rocm page very specifically does not mention any other steps in the Arch section and the AUR page only mentions the UI issue.



  • I distrohopped so much after each previous distro eventually broke and me clearly not being smart enough to recover. I’m honestly kinda sick of it, even if the immutable nature also annoys the shit out of me.

    My GPU is a 6650 XT, which should in principle work with ROCm.

    Which model specifically are you recommending? Llama-3.1-8B-Lexi-Uncensored-V2-GGUF? Because the original meta-llama ones are censored to all hell and Huggingface is not particularly easy to navigate, on top of figuring out the right model size & quantization being extremely confusing.


  • I just can’t get ROCm / gpu generation to work on Bazzite, like at all. It seems completely cursed. I tried koboldcpp through a Fedora distrobox and it didn’t even show any hardware options. Tried through an Arch AUR package through distrobox and the ROCm option is there but ends with a CUDA error. lol The Vulkan option works but seems to still use the CPU more than the GPU and is consequently still kinda slow and I struggle to find a good model for my 8GB card. Fimbulvetr-10.7B-v1-Q5_K_M for example was still too slow to be practical.

    Tried LM Studio directly in Bazzite and it also just uses the CPU. It also is very obtuse on how to connect to it with SillyTavern, as it asks for an API key? I managed it once in the past but I can’t remember how but it also ended up stopping generating anything after a few replies.

    Krita’s diffusion also only runs on the CPU, which is abysmally slow, but I’m not sure if they expect Krita to be build directly on the system for ROCm support to work.

    I’m not even trying to get SDXL or something to run at this point, since that seems to be still complicated enough even on a regular distro.