NixOS CUDA | 284 Members | |
| CUDA packages maintenance and support in nixpkgs | https://github.com/orgs/NixOS/projects/27/ | https://nixos.org/manual/nixpkgs/unstable/#cuda | 58 Servers |
| Sender | Message | Time |
|---|---|---|
| 16 Mar 2023 | ||
| We should now be able to run nixpkgs-review like so: https://github.com/Mic92/nixpkgs-review/pull/315 | 18:12:15 | |
| Someone S: is there a way to see a history of your CI runs that get pushed to cuda-maintainers? i'm curious to see what things are building and what aren't atm | 23:15:25 | |
| i seem to be getting a bunch of cache misses in my CI that i would not expect | 23:15:30 | |
| Samuel Ainsworth: oh, this is almost certainly happening because I decreased the frequency for the default set of
You can see which jobs sets are enabled and what schedule they run on in any "config" job in hercules, e.g.: https://hercules-ci.com/github/SomeoneSerge/nixpkgs-unfree/jobs/3513 | 23:26:17 | |
Sorry I did that without proper announcement | 23:26:39 | |
| This reminds me, I wanted to open https://github.com/NixOS/nixpkgs/issues/221564 | 23:28:40 | |
| 17 Mar 2023 | ||
| 04:31:41 | ||
| Are people mostly deploying torch/cuda builds via nix-shells nowadays? I've had bad luck getting poetry2nix/dream2nix working for cuda enabled torch builds and was curious what the workflow is with the cuda cachix now | 04:44:09 | |
| I can only speak for myself, I use source builds from nixpkgs with cudaSupport=true. What kind of issues are we talking about? | 11:04:56 | |
| we also have cuda support for devenv, but it's not finalized yet: https://github.com/cachix/devenv/pull/422 | 12:24:56 | |
| how does devenv work to access the host graphics drivers? | 13:33:11 | |
I ended up using a shellHook which aliases python and other common commands so they're wrapped with nixGL: https://github.com/ConnorBaker/mfsr_utils/blob/main/flake.nix#L82-L86 | 13:44:55 | |
| I have some time today so hopefully I can get caught up on all the stuff merged/reported over the last two weeks ✨ | 13:47:29 | |
| Per the conversation here https://github.com/NixOS/nixpkgs/pull/220366#discussion_r1135048161, what needs to be done to tell the community that we're dropping support for CUDA 10? Does it make sense to stop there, or should we drop up until 11.4? In terms of the GPUs we support, the only thing we'd remove support for is the early-generation Kepler GPUs: https://github.com/NixOS/nixpkgs/blob/master/pkgs/development/compilers/cudatoolkit/gpus.nix#L18-L31, since everything else is supported through at least 11.8. | 14:36:36 | |
In reply to @connorbaker:matrix.orgI don't think we should remove CUDA 10 as such, until it's officially EOL (is it?). We should just communicate that when we package e.g. pytorch in nixpkgs we'll ensure that it works with cudaPackages_11_4 forward, but if one wants to build against cudaPackages_10, they'll just have to maintain their own copy of the torch expression | 14:43:53 | |
I think this amounts to a message in the release notes: non-legacy packages can be overridden and passed a different cudaPackages argument, as long as it's redist | 14:45:11 | |
| Hm, I realize I don't actually know what kind of support NVIDIA provides for different devices. I know on Windows they'll remove support for devices by model, but with their Linux drivers it's usually per-architecture. I did find https://docs.nvidia.com/datacenter/tesla/drivers/#software-matrix, but I don't know if they define EOL/unsupported/deprecated For example, here's them announcing the end of support for Quadro Kepler devices: https://nvidia.custhelp.com/app/answers/detail/a_id/5210 | 14:56:13 | |
| Maybe this is made more difficult by the fact that there's driver support and then there's software library (CUDA) support :l I guess it'd be fair to say that if they're no longer publishing drivers for a GPU that it has been EOL'd? | 14:58:03 | |
Lol. Maybe when the last package in nixpkgs drops the cuda < 11 requirement | 15:08:38 | |
| ...which means what, caffe? | 15:08:54 | |
| I can send a couple repros in the channel! | 15:50:33 | |
| For the nix-shell case, here is a minimal example: | 15:50:53 | |
| 15:51:56 | |
| 15:52:14 | |
| This doesn't seem to hit the binary cache? | 15:52:18 | |
| Thanks Domen! I'm going to try out the cuda support with the poetry example https://github.com/cachix/devenv/tree/main/examples/python-poetry once it's merged | 15:57:37 | |
| Thanks! This is how I used to do things, but one issue is that it's locked to whatever the latest is in nixpkgs. Btw, how do you avoid cache misses? Is there a list of what cuda-maintainers is providing CI for? | 16:05:13 | |
| Ahh I see why you had to do the nvidia driver pinning, thats a bit unfortunate | 16:26:21 | |
| Is there a reason you didn't opt for setting the LD_LIBRARY_PATH directly? | 16:34:41 | |
| 16:34:45 | |