I am on an headless server Ubuntu 20.04 and I have an NVIDIA Tesla P4, for which I have installed NVIDIA drivers 525 with CUDA 12.0 from the NVIDIA website (.run, not with ubuntu-drivers)
It seems like the server sees the GPU fine:
Code:
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 525.147.05 Driver Version: 525.147.05 CUDA Version: 12.0 |
|-------------------------------+----------------------+----------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|===============================+======================+======================|
| 0 Tesla P4 Off | 00000000:21:00.0 Off | 0 |
| N/A 40C P0 22W / 75W | 0MiB / 7680MiB | 0% Default |
| | | N/A |
+-------------------------------+----------------------+----------------------+
+-----------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=============================================================================|
| No running processes found |
+-----------------------------------------------------------------------------+
Code:
inxi -G
Graphics: Device-1: NVIDIA GP104GL [Tesla P4] driver: nvidia v: 525.147.05
Display: server: X.Org 1.21.1.6 driver: nvidia resolution: 4080x2160~1Hz
OpenGL: renderer: Tesla P4/PCIe/SSE2 v: 4.6.0 NVIDIA 525.147.05
Code:
lspci -v | grep -EA3 'VGA|3D'
01:00.1 VGA compatible controller: Matrox Electronics Systems Ltd. MGA G200EH (prog-if 00 [VGA controller])
Subsystem: Hewlett-Packard Company iLO4
Flags: bus master, fast devsel, latency 0, IRQ 16, NUMA node 0
Memory at d9000000 (32-bit, prefetchable) [size=16M]
--
21:00.0 3D controller: NVIDIA Corporation GP104GL [Tesla P4] (rev a1)
Subsystem: NVIDIA Corporation GP104GL [Tesla P4]
Physical Slot: 4
Flags: bus master, fast devsel, latency 0, IRQ 110, NUMA node 1
however when I tell it to use nvidia, it gives me this:
Code:
sudo prime-select nvidia
Error: no integrated GPU detected.
Or when I try to use the GPU with software like llama.cpp, trying to load a LLM to the GPU, it ends up using only the CPU only as if I had no GPU...
my end goal is also to use the GPU in X-forwarding to run programs that need to build OpenGL windows. If I run
I see the XQuartz window popping up, but any other command like
Code:
glxgears
Error: couldn't get an RGB, Double-buffered visual
or
Code:
glmark2
Error: glXChooseFBConfig() failed
Error: Error: Couldn't get GL visual config!
Error: main: Could not initialize canvas
So I guess first thing I want to be sure that the server sees the GPU... are there any other test I can try to see if actually my server sees the GPU? or to select it definitively as the de-facto video card?