forked from tabletuser-blogspot/ollama-benchmark
-
Notifications
You must be signed in to change notification settings - Fork 14
Open
Description
See: geerlingguy/sbc-reviews#92
To get llama.cpp to compile with CUDA, I had to specify the CUDA compiler path:
cmake -B build -DGGML_CUDA=1 -DCMAKE_CUDA_COMPILER=/usr/local/cuda/bin/nvccI also got the error:
ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 CUDA devices:
Device 0: NVIDIA GB10, compute capability 12.1, VMM: yes
MESA: error: Opening /dev/dri/card0 failed: Permission denied
TU: error: ../src/freedreno/vulkan/tu_knl.cc:385: failed to open device /dev/dri/renderD128 (VK_ERROR_INCOMPATIBLE_DRIVER)
ggml_vulkan: Found 1 Vulkan devices:
ggml_vulkan: 0 = NVIDIA Tegra NVIDIA GB10 (NVIDIA) | uma: 1 | fp16: 1 | bf16: 0 | warp size: 32 | shared memory: 49152 | int dot: 0 | matrix cores: KHR_coopmat
...
So I added my user account to the video and render groups:
sudo usermod -aG video,render jgeerling
Metadata
Metadata
Assignees
Labels
No labels