6 Jul 2024 |
| @anjannath:matrix.org left the room. | 04:50:52 |
| jeroenvb3 joined the room. | 23:43:32 |
jeroenvb3 | Hi there, could someone point me to up to date documentation for getting python to work with cuda? I have tried some things but I don't know what the best/official way is that I should start with. Thanks. | 23:54:01 |
7 Jul 2024 |
Someone (UTC+3) | import <nixpkgs> { config.cudaSupport = true } | 00:59:29 |
| @1h0:matrix.org left the room. | 08:53:13 |
jeroenvb3 | Could you please give me a bit more information? I added it to a shell.nix which seemed promising:
# Run with `nix-shell cuda-shell.nix`
# { pkgs ? import <nixpkgs> {} }:
with import <nixpkgs> {
config = {
allowUnfree = true;
cudaSupport = true;
};
};
pkgs.mkShell {
name = "cuda-env-shell";
buildInputs = with pkgs; [
git gitRepo gnupg autoconf curl
procps gnumake util-linux m4 gperf unzip
cudatoolkit linuxPackages.nvidia_x11
libGLU libGL
xorg.libXi xorg.libXmu freeglut
xorg.libXext xorg.libX11 xorg.libXv xorg.libXrandr zlib
ncurses5 stdenv.cc binutils
python39
python39Packages.numpy
python39Packages.numba
libstdcxx5
cudaPackages_11.cudatoolkit
];
shellHook = ''
export CUDA_PATH=${pkgs.cudatoolkit}
# export LD_LIBRARY_PATH=${pkgs.linuxPackages.nvidia_x11}/lib:${pkgs.ncurses5}/lib
export EXTRA_LDFLAGS="-L/lib -L${pkgs.linuxPackages.nvidia_x11}/lib"
export EXTRA_CCFLAGS="-I/usr/include"
export LD_LIBRARY_PATH=${pkgs.cudaPackages_11.cudatoolkit}/lib:$LD_LIBRARY_PATH
export NUMBAPRO_NVVM=${pkgs.cudaPackages_11.cudatoolkit}/nvvm/lib64/libnvvm.so
export NUMBAPRO_LIBDEVICE=${pkgs.cudaPackages_11.cudatoolkit}/nvvm/libdevice
'';
}
so when I go into that shell, create a python virtualenv, go into it, install numba. And try to run:
from numba import cuda
print(cuda.detect())
It gives me:
Numba Version: 0.59.1
Traceback (most recent call last):
File "/tmp/cuda/test.py", line 39, in <module>
main()
File "/tmp/cuda/test.py", line 19, in main
print_version_info()
File "/tmp/cuda/test.py", line 7, in print_version_info
print("CUDA Version:", cuda.runtime.get_version())
File "/nix/store/7m7c6crkdbzmzcrbwa4l4jqgnwj8m92b-python3.9-numba-0.59.1/lib/python3.9/site-packages/numba/cuda/cudadrv/runtime.py", line 111, in get_version
self.cudaRuntimeGetVersion(ctypes.byref(rtver))
File "/nix/store/7m7c6crkdbzmzcrbwa4l4jqgnwj8m92b-python3.9-numba-0.59.1/lib/python3.9/site-packages/numba/cuda/cudadrv/runtime.py", line 81, in safe_cuda_api_call
self._check_error(fname, retcode)
File "/nix/store/7m7c6crkdbzmzcrbwa4l4jqgnwj8m92b-python3.9-numba-0.59.1/lib/python3.9/site-packages/numba/cuda/cudadrv/runtime.py", line 89, in _check_error
raise CudaRuntimeAPIError(retcode, msg)
numba.cuda.cudadrv.runtime.CudaRuntimeAPIError: [34] Call to cudaRuntimeGetVersion results in CUDA_ERROR_STUB_LIBRARY
I would not mind reading more, but I can't find good sources for nixos cuda support to jump off from. Thanks again.
| 09:28:09 |
Someone (UTC+3) | In reply to @jeroenvb3:matrix.org
Could you please give me a bit more information? I added it to a shell.nix which seemed promising:
# Run with `nix-shell cuda-shell.nix`
# { pkgs ? import <nixpkgs> {} }:
with import <nixpkgs> {
config = {
allowUnfree = true;
cudaSupport = true;
};
};
pkgs.mkShell {
name = "cuda-env-shell";
buildInputs = with pkgs; [
git gitRepo gnupg autoconf curl
procps gnumake util-linux m4 gperf unzip
cudatoolkit linuxPackages.nvidia_x11
libGLU libGL
xorg.libXi xorg.libXmu freeglut
xorg.libXext xorg.libX11 xorg.libXv xorg.libXrandr zlib
ncurses5 stdenv.cc binutils
python39
python39Packages.numpy
python39Packages.numba
libstdcxx5
cudaPackages_11.cudatoolkit
];
shellHook = ''
export CUDA_PATH=${pkgs.cudatoolkit}
# export LD_LIBRARY_PATH=${pkgs.linuxPackages.nvidia_x11}/lib:${pkgs.ncurses5}/lib
export EXTRA_LDFLAGS="-L/lib -L${pkgs.linuxPackages.nvidia_x11}/lib"
export EXTRA_CCFLAGS="-I/usr/include"
export LD_LIBRARY_PATH=${pkgs.cudaPackages_11.cudatoolkit}/lib:$LD_LIBRARY_PATH
export NUMBAPRO_NVVM=${pkgs.cudaPackages_11.cudatoolkit}/nvvm/lib64/libnvvm.so
export NUMBAPRO_LIBDEVICE=${pkgs.cudaPackages_11.cudatoolkit}/nvvm/libdevice
'';
}
so when I go into that shell, create a python virtualenv, go into it, install numba. And try to run:
from numba import cuda
print(cuda.detect())
It gives me:
Numba Version: 0.59.1
Traceback (most recent call last):
File "/tmp/cuda/test.py", line 39, in <module>
main()
File "/tmp/cuda/test.py", line 19, in main
print_version_info()
File "/tmp/cuda/test.py", line 7, in print_version_info
print("CUDA Version:", cuda.runtime.get_version())
File "/nix/store/7m7c6crkdbzmzcrbwa4l4jqgnwj8m92b-python3.9-numba-0.59.1/lib/python3.9/site-packages/numba/cuda/cudadrv/runtime.py", line 111, in get_version
self.cudaRuntimeGetVersion(ctypes.byref(rtver))
File "/nix/store/7m7c6crkdbzmzcrbwa4l4jqgnwj8m92b-python3.9-numba-0.59.1/lib/python3.9/site-packages/numba/cuda/cudadrv/runtime.py", line 81, in safe_cuda_api_call
self._check_error(fname, retcode)
File "/nix/store/7m7c6crkdbzmzcrbwa4l4jqgnwj8m92b-python3.9-numba-0.59.1/lib/python3.9/site-packages/numba/cuda/cudadrv/runtime.py", line 89, in _check_error
raise CudaRuntimeAPIError(retcode, msg)
numba.cuda.cudadrv.runtime.CudaRuntimeAPIError: [34] Call to cudaRuntimeGetVersion results in CUDA_ERROR_STUB_LIBRARY
I would not mind reading more, but I can't find good sources for nixos cuda support to jump off from. Thanks again.
In this instance you have two nixpkgs instance, pkgs without cuda support, and another implicit one in the top-level with expression. In buildInputs = with pkgs; [ ... ] you're still using packages from the no-cuda instance | 09:33:10 |
Someone (UTC+3) | Note that you don't need cudatoolkit in mkShells unless you're setting up a c++/cuda dev environment | 09:33:49 |
Someone (UTC+3) | You don't need linuxPackages.nvidia_x11 in mkShells, it'll only break things | 09:34:31 |
Someone (UTC+3) |
numba.cuda.cudadrv.runtime.CudaRuntimeAPIError: [34] Call to cudaRuntimeGetVersion results in CUDA_ERROR_STUB_LIBRARY
COmes from a symlink to the stub library in ${cudatoolkit}/lib/libcuda.so which you've listed in LD_LIBRARY_PATH and thus gave it a higher priority than the path to the real driver recorded in libraries' and executables' headers already | 09:35:59 |
Someone (UTC+3) | Note: the symlink at ${cudatoolkit}/lib/libcuda.so was removed in a recent PR and this particular error will go away once it reaches nixos-unstable. NOnetheless, you don't usually need cudatoolkit in LD_LIBRARY_PATH | 09:37:11 |
| @dminca:matrix.org joined the room. | 09:37:45 |
Someone (UTC+3) | * In this instance you have two nixpkgs instances, pkgs without cuda support, and another implicit one in the top-level with expression. In buildInputs = with pkgs; [ ... ] you're still using packages from the no-cuda instance | 09:38:59 |
jeroenvb3 | Thank you very much. The previously mentioned file does indeed run succesfully now. I did want to enable cudatoolkit I'm pretty sure. This is what I am now trying to get running:
from numba import cuda
import numpy as np
@cuda.jit
def cudakernel0(array):
for i in range(array.size):
array[i] += 0.5
array = np.array([0, 1], np.float32)
print('Initial array:', array)
print('Kernel launch: cudakernel0[1, 1](array)')
cudakernel0[1, 1](array)
print('Updated array:',array)
Which has this as the first error:
Initial array: [0. 1.]
Kernel launch: cudakernel0[1, 1](array)
/nix/store/7m7c6crkdbzmzcrbwa4l4jqgnwj8m92b-python3.9-numba-0.59.1/lib/python3.9/site-packages/numba/cuda/dispatcher.py:536: NumbaPerformanceWarning: Grid size 1 will likely result in GPU under-utilization due to low occupancy.
warn(NumbaPerformanceWarning(msg))
Traceback (most recent call last):
File "/nix/store/7m7c6crkdbzmzcrbwa4l4jqgnwj8m92b-python3.9-numba-0.59.1/lib/python3.9/site-packages/numba/cuda/cudadrv/nvvm.py", line 139, in __new__
inst.driver = open_cudalib('nvvm')
File "/nix/store/7m7c6crkdbzmzcrbwa4l4jqgnwj8m92b-python3.9-numba-0.59.1/lib/python3.9/site-packages/numba/cuda/cudadrv/libs.py", line 64, in open_cudalib
return ctypes.CDLL(path)
File "/nix/store/2j0l3b15gas78h9akrsfyx79q02i46hc-python3-3.9.19/lib/python3.9/ctypes/__init__.py", line 374, in __init__
self._handle = _dlopen(self._name, mode)
OSError: libnvvm.so: cannot open shared object file: No such file or directory
However, I do have the related env vars set:
[nix-shell:/tmp/cuda]$ echo $NUMBAPRO_NVVM
/nix/store/0c8nf26hx9x9jxgj0s9bq10xg75nbfv0-cuda-merged-12.2/nvvm/lib64/libnvvm.so
[nix-shell:/tmp/cuda]$ echo $NUMBAPRO_LIBDEVICE
/nix/store/0c8nf26hx9x9jxgj0s9bq10xg75nbfv0-cuda-merged-12.2/nvvm/libdevice
A stackoverflow says those are outdated and CUDA_HOME needs to be set. I do set it to the same as CUDA_PATH, but it doesn't seem to help. http://numba.pydata.org/numba-doc/latest/cuda/overview.html#setting-cuda-installation-path talks about ignoring minor version paths, but I don't think that is when I directly set it. I also don't have any non-minor version paths in /nix/store/
This is now my shell.nix:
with import <nixpkgs> {
config = {
allowUnfree = true;
cudaSupport = true;
};
};
pkgs.mkShell {
name = "cuda-env-shell";
buildInputs = [
git gitRepo gnupg autoconf curl
procps gnumake util-linux m4 gperf unzip
libGLU libGL
xorg.libXi xorg.libXmu freeglut
xorg.libXext xorg.libX11 xorg.libXv xorg.libXrandr zlib
ncurses5 stdenv.cc binutils
python39
python39Packages.numpy
python39Packages.numba
libstdcxx5
cudaPackages.cudatoolkit
];
shellHook = ''
export CUDA_PATH=${pkgs.cudaPackages.cudatoolkit}
export CUDA_HOME=${pkgs.cudatoolkit}
export EXTRA_CCFLAGS="-I/usr/include"
export NUMBAPRO_NVVM=${pkgs.cudatoolkit}/nvvm/lib64/libnvvm.so
export NUMBAPRO_LIBDEVICE=${pkgs.cudatoolkit}/nvvm/libdevice
'';
}
I'm sorry if its a lot to ask, but I would really like to learn about this and get it working. Do you see anything wrong with my stuff still?
| 15:29:13 |
| @wilhelmvon:matrix.org left the room. | 18:36:51 |
8 Jul 2024 |
Someone (UTC+3) |
CUDA_PATH, but it doesn't seem to help. http://numba.pydata.org/numba-doc/latest/cuda/overview.html#setting-cuda-installation-path talks
Hmm are those used at build time or at runtime for jit? | 16:05:39 |
Someone (UTC+3) | In reply to @jeroenvb3:matrix.org
Thank you very much. The previously mentioned file does indeed run succesfully now. I did want to enable cudatoolkit I'm pretty sure. This is what I am now trying to get running:
from numba import cuda
import numpy as np
@cuda.jit
def cudakernel0(array):
for i in range(array.size):
array[i] += 0.5
array = np.array([0, 1], np.float32)
print('Initial array:', array)
print('Kernel launch: cudakernel0[1, 1](array)')
cudakernel0[1, 1](array)
print('Updated array:',array)
Which has this as the first error:
Initial array: [0. 1.]
Kernel launch: cudakernel0[1, 1](array)
/nix/store/7m7c6crkdbzmzcrbwa4l4jqgnwj8m92b-python3.9-numba-0.59.1/lib/python3.9/site-packages/numba/cuda/dispatcher.py:536: NumbaPerformanceWarning: Grid size 1 will likely result in GPU under-utilization due to low occupancy.
warn(NumbaPerformanceWarning(msg))
Traceback (most recent call last):
File "/nix/store/7m7c6crkdbzmzcrbwa4l4jqgnwj8m92b-python3.9-numba-0.59.1/lib/python3.9/site-packages/numba/cuda/cudadrv/nvvm.py", line 139, in __new__
inst.driver = open_cudalib('nvvm')
File "/nix/store/7m7c6crkdbzmzcrbwa4l4jqgnwj8m92b-python3.9-numba-0.59.1/lib/python3.9/site-packages/numba/cuda/cudadrv/libs.py", line 64, in open_cudalib
return ctypes.CDLL(path)
File "/nix/store/2j0l3b15gas78h9akrsfyx79q02i46hc-python3-3.9.19/lib/python3.9/ctypes/__init__.py", line 374, in __init__
self._handle = _dlopen(self._name, mode)
OSError: libnvvm.so: cannot open shared object file: No such file or directory
However, I do have the related env vars set:
[nix-shell:/tmp/cuda]$ echo $NUMBAPRO_NVVM
/nix/store/0c8nf26hx9x9jxgj0s9bq10xg75nbfv0-cuda-merged-12.2/nvvm/lib64/libnvvm.so
[nix-shell:/tmp/cuda]$ echo $NUMBAPRO_LIBDEVICE
/nix/store/0c8nf26hx9x9jxgj0s9bq10xg75nbfv0-cuda-merged-12.2/nvvm/libdevice
A stackoverflow says those are outdated and CUDA_HOME needs to be set. I do set it to the same as CUDA_PATH, but it doesn't seem to help. http://numba.pydata.org/numba-doc/latest/cuda/overview.html#setting-cuda-installation-path talks about ignoring minor version paths, but I don't think that is when I directly set it. I also don't have any non-minor version paths in /nix/store/
This is now my shell.nix:
with import <nixpkgs> {
config = {
allowUnfree = true;
cudaSupport = true;
};
};
pkgs.mkShell {
name = "cuda-env-shell";
buildInputs = [
git gitRepo gnupg autoconf curl
procps gnumake util-linux m4 gperf unzip
libGLU libGL
xorg.libXi xorg.libXmu freeglut
xorg.libXext xorg.libX11 xorg.libXv xorg.libXrandr zlib
ncurses5 stdenv.cc binutils
python39
python39Packages.numpy
python39Packages.numba
libstdcxx5
cudaPackages.cudatoolkit
];
shellHook = ''
export CUDA_PATH=${pkgs.cudaPackages.cudatoolkit}
export CUDA_HOME=${pkgs.cudatoolkit}
export EXTRA_CCFLAGS="-I/usr/include"
export NUMBAPRO_NVVM=${pkgs.cudatoolkit}/nvvm/lib64/libnvvm.so
export NUMBAPRO_LIBDEVICE=${pkgs.cudatoolkit}/nvvm/libdevice
'';
}
I'm sorry if its a lot to ask, but I would really like to learn about this and get it working. Do you see anything wrong with my stuff still?
AFAIU it dlopen() s with the soname rather than an absolute path, which is probably good for us | 16:06:41 |
Someone (UTC+3) | You could try exposing libnvvm in LD_LIBRARY_PATH | 16:06:49 |
Someone (UTC+3) | It would be valuable if we managed to prepare runtime tests for numba/numba+cuda | 16:07:51 |
Someone (UTC+3) |
Do you see anything wrong with my stuff still?
I'm still unsure if all of the buildInputs are actually required for whatever your task is. Aside from that, it's probably best that you create a python wrapper (python3.withPackages (ps: with ps; [ numba ... ]) ) rather than rely on mkShell or whatever | 16:09:39 |
Someone (UTC+3) | If you build cudaPackages.cudatoolkit you'll see that it's actually a symlink farm which includes a lot of stuff you don't need. In particular:
❯ readlink -f result/nvvm/lib64/libnvvm.so.4.0.0
/nix/store/fby2d6b4jgfb8awwjhzdrd13r8vx7ilw-cuda_nvcc-12.2.140-bin/nvvm/lib64/libnvvm.so.4.0.0
...it would've been enough for you to use ${getBin cudaPackages.cuda_nvcc}/nvvm/lib64/libnvvm.so which is much smaller (whoops, that's a bug, we definitely didn't want this to end up in .bin , so expect this to change) | 16:12:09 |
9 Jul 2024 |
| ghishadow joined the room. | 04:20:35 |
jeroenvb3 | I would think runtime for jit, since there is a @cuda.jit operator on my python code. This is now my shell.nix, removed extra buildInputs, and used the wrapper, which had to be python39 to even get back to where I was:
config = {
allowUnfree = true;
cudaSupport = true;
};
};
let
pythonEnv = pkgs.python39.withPackages(ps: with ps; [ numba numpy ]);
in
pkgs.mkShell {
name = "cuda-env-shell";
buildInputs = [
pythonEnv
libstdcxx5
cudaPackages.cudatoolkit
];
shellHook = ''
export CUDA_PATH=${pkgs.cudaPackages.cudatoolkit}
export CUDA_HOME=${pkgs.cudatoolkit}
export EXTRA_CCFLAGS="-I/usr/include"
export NUMBAPRO_NVVM=${pkgs.cudatoolkit}/nvvm/lib64/libnvvm.so
export NUMBAPRO_LIBDEVICE=${pkgs.cudatoolkit}/nvvm/libdevice
export LD_LIBRARY_PATH=${pkgs.cudatoolkit}/nvvm/lib64/libnvvm.so:$CUDA_HOME
# source venv/bin/activate
python test2.py
'';
}
If I go into python repl I can import numpy and numba, but I still get the same error when running an actual cuda script. This was the python code:
from numba import cuda
import numpy as np
@cuda.jit
def cudakernel0(array):
for i in range(array.size):
array[i] += 0.5
array = np.array([0, 1], np.float32)
print('Initial array:', array)
print('Kernel launch: cudakernel0[1, 1](array)')
cudakernel0[1, 1](array)
print('Updated array:',array)
I can run:
from numba import cuda
print(cuda.detect())
Also I doubt the linking of nvvm.so helps, since numba is now relying on the CUDA_HOME env var (https://numba.pydata.org/numba-doc/latest/cuda/overview.html). What do you mean to use ${getBin...}, where could I use that?
| 10:05:31 |
Someone (UTC+3) |
export LD_LIBRARY_PATH=${pkgs.cudatoolkit}/nvvm/lib64/libnvvm.so:$CUDA_HOME
I'm pretty sure gilbc treats LD_LIBRARY_PATH as a list of directories only (so you'd want LD_LIBRARY_PATH=${pkgs.cudatoolkit}/nvvm/lib64/ instead) | 10:42:21 |
Someone (UTC+3) |
Also I doubt the linking of nvvm.so helps, since numba is now relying on the CUDA_HOME env var
The error you posted earlier suggests that they just dlopen("libnvvm.so", ...) | 10:42:57 |
jeroenvb3 | In reply to @ss:someonex.net
export LD_LIBRARY_PATH=${pkgs.cudatoolkit}/nvvm/lib64/libnvvm.so:$CUDA_HOME
I'm pretty sure gilbc treats LD_LIBRARY_PATH as a list of directories only (so you'd want LD_LIBRARY_PATH=${pkgs.cudatoolkit}/nvvm/lib64/ instead) That was true yes, it seems it does find it now. However it claims:
No supported GPU compute capabilities found. Please check your cudatoolkit version matches your CUDA version
nvcc --version gives 12.2, and nvidia-smi gives 12.4. This should be compatible as far as I'm aware. I'll see if I can get more info on this, unless you know that this is not supposed to be compatible and I need to adjust one of the versions.
| 12:26:13 |
Someone (UTC+3) | In reply to @jeroenvb3:matrix.org
That was true yes, it seems it does find it now. However it claims:
No supported GPU compute capabilities found. Please check your cudatoolkit version matches your CUDA version
nvcc --version gives 12.2, and nvidia-smi gives 12.4. This should be compatible as far as I'm aware. I'll see if I can get more info on this, unless you know that this is not supposed to be compatible and I need to adjust one of the versions.
Could you gist a reproducing code and a full log? | 12:30:02 |
10 Jul 2024 |
jeroenvb3 | In reply to @ss:someonex.net Could you gist a reproducing code and a full log? https://gitlab.com/jeroenvb3/cuda-setup
I added the nix config, the python code, the output, and my nvidia driver info. If you need more system info please tell but I'd think this is all that will be unique. Thanks!
| 10:23:54 |
12 Jul 2024 |
| valconius left the room. | 01:16:43 |
15 Jul 2024 |
| @hasnep:matrix.org joined the room. | 01:58:17 |