NixOS CUDA | 286 Members | |
| CUDA packages maintenance and support in nixpkgs | https://github.com/orgs/NixOS/projects/27/ | https://nixos.org/manual/nixpkgs/unstable/#cuda | 58 Servers |
| Sender | Message | Time |
|---|---|---|
| 10 Sep 2025 | ||
In reply to @hugo:okeso.euAt the very least we recently stopped early-binding rocm libraries, maybe hard-coded compiler paths went with them. Try giving it a compiler at test time as suggested in the error? | 10:05:31 | |
| I (vibe) tried to give it a compiler here in this commit https://github.com/NixOS/nixpkgs/pull/441728/commits/81f7997ca1ca37193f2f26fdbc85c586a92ba6dd but was unsuccessful. Any suggestion how to do that? | 10:06:58 | |
| 15:02:50 | ||
| That should only impact ROCm unless I messed it up! diff was https://github.com/NixOS/nixpkgs/commit/c74e5ffb6526ac1b4870504921b9ba9362189a17 | 15:52:00 | |
| 18:26:24 | ||
| Is this team involved in flox/nvidia partnership ? (See https://flox.dev/cuda/) I guess so since the nixos foundation also is, but there is no mention of this team or its amazing work. | 18:30:16 | |
| 19:53:26 | |
| After a whole day recompiling torch :) | 19:53:32 | |
| Actually, with the HSA override to 11.0.0 it worked, but I get a different kind of error
| 19:58:53 | |
| * Actually, with the HSA override to 11.0.0 it worked, but I get a different kind of error
| 19:59:14 | |
In reply to @layus:matrix.orgRon mentioned them here https://discourse.nixos.org/t/nix-flox-nvidia-opening-up-cuda-redistribution-on-nix/69189/7 | 20:01:06 | |
| Is there a rocm room? | 20:54:56 | |
| https://matrix.to/#/#ROCm:nixos.org | 21:40:38 | |
| 22:20:37 | ||
| Well done guys for allowing this to happen (connor (he/him) (UTC-7) SomeoneSerge (back on matrix) stick...) 👏 | 23:07:17 | |
| * Well done guys for allowing this to happen (connor (he/him) (UTC-7) SomeoneSerge (back on matrix) stick Samuel Ainsworth...) 👏 | 23:22:06 | |
| The negotiations with NVIDIA have been run by Flox (although in parallel with many other companies' simultaneous inquiries). Ron kept us, the Foundation, and the SC in the loop, and offered both legal help and workforce. The current idea roughly is that the CUDA team gets access to the relevant repo and infra, and work closely together with Flox to secure the position and a commx channel to NVIDIA. | 23:26:05 | |
| What were the blockers for setting this up within the NixOS Foundation? | 23:54:32 | |
| 11 Sep 2025 | ||
| From what I recall, it was something to do with having a legal entity in the US. If the foundation was registered in the US, then it would've been fine. This has been going on since at least Planet Nix, glad to see it finally pull through. | 00:03:41 | |
| 00:09:51 | ||
| Will try to take a look later | 00:10:55 | |
| You’d probably need to override writeGpuTestPython to use the Python package set you specify. It’s inside cudaPackages so it has no way of knowing what scope you’re using it in. | 00:12:26 | |
| They’re still mostly okay; I’ve been exhausted recently so haven’t been awake too early (I was up at 3am this morning but that’s something else) If we can do something closer to 8 that might be easier; Kevin Mittman how are you with morning meetings? It’d be good to catch up and discuss what’s been done so far with the database project Serge’s been working on. | 00:14:16 | |
| NVIDIA’s EULA effectively prohibits running CUDA binaries they release on non-NVIDIA hardware (see 1.2.8: https://docs.nvidia.com/cuda/eula/index.html#limitations) | 00:20:07 | |
| https://developer.nvidia.com/blog/developers-can-now-get-cuda-directly-from-their-favorite-third-party-platforms/ | 01:17:14 | |
| The fact remains: it was amd who shut it down, not nvidia? | 01:26:18 | |
| nvidia playing the "please submit this in paper by post, and attach proofs of your residence such as electricity bills delivered to your address" game (being the bureaucrat and coming up with arbitrary terms as they go) | 01:30:19 | |
| ah yeah, true | 01:33:21 | |
| When is the next one? I'll try to join | 10:16:38 | |
I've updated the gist with the latest version, it seems to compile and run a pipeline with tensor_filter framework=pytorch accelerator=true:gpu ... ! fakesink, but I haven't got a time to check it really. | 10:45:16 | |