Replies: 3 comments
-
I upgraded cmake to version 3.31.2 and I no longer get the error: I can compile/run but GPUs are still not detected at runtime. |
Beta Was this translation helpful? Give feedback.
-
If it's any help, I'm also at this step but I've tested other things (e.g. pytorch) and while the GPUs in the system show up in Our system is an 8x PCIe H200 NVL system. e..g pytorch
|
Beta Was this translation helpful? Give feedback.
-
After a lot of swearing, I build the cuda-samples deviceQuery sample, and it output this:
Googling that error suggested down/side-grading to the proprietary driver from the Open one which I did, rebooted and now things mostly work (I still have a problem in torch with multiple GPUs but I suspect that's a differnt problem entirely) - so I suggest you try that? |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I'm having problems building llama.cpp on a DGX-H200.
Driver Version: 570.158.01 CUDA Version: 12.8
cmake version is 3.26.5
nvcc is in the normal /usr/local/cuda and is detected
running cmake -B build -DGGML_CUDA=1 ends with:
CMake Error in ggml/src/ggml-cuda/CMakeLists.txt:
CUDA_ARCHITECTURES is set to "native", but no GPU was detected.
If I overrride and build using:
cmake3 -B build -DGGML_CUDA=1 -DCMAKE_CUDA_ARCHITECTURES=89
It builds without error but when executed does not use the GPU. I suspect there is something about our system that is non-standard that is keeping llama.cpp from detecting the GPUs at build time (native) and at run time. Is there something I overlooked in the configuration?
Beta Was this translation helpful? Give feedback.
All reactions