| 11 Oct 2025 |
SomeoneSerge (back on matrix) | In reply to @rosscomputerguy:matrix.org Hey, connor (he/him) (UTC-7) & SomeoneSerge (back on matrix). Either of you wanna collab on getting Tenstorrent support into nixpkgs? I'm the only one working on it but I think since this is in a realm of AI, ML, and GPU-like computing, it would make sense to involve people already touching that stuff. YES! /looks at the calendar, lowers the volume/ yes, though very much part time god hiw do i learn to say no | 18:17:28 |
SomeoneSerge (back on matrix) | * YES! /looks at the calendar, lowers the volume/ yes, though very much part time god how do i learn to say no | 18:17:37 |
| 12 Oct 2025 |
Tristan Ross | Heh, it's not too much. Bulk of the reviewing is https://github.com/NixOS/nixpkgs/pull/444813 | 17:19:19 |
| 13 Oct 2025 |
Collin Arnett | Hello! Have ya'll run into this problem with the nvidia-runtime-container?
https://github.com/llm-d/llm-d/issues/117#issuecomment-2992256350
apparently there is a patch for it here https://github.com/NVIDIA/k8s-device-plugin/pull/1183/files | 15:41:41 |
Collin Arnett | * Hello! Have ya'll run into this problem with the nvidia-container-toolkit?
https://github.com/llm-d/llm-d/issues/117#issuecomment-2992256350
apparently there is a patch for it here https://github.com/NVIDIA/k8s-device-plugin/pull/1183/files | 15:52:49 |
SomeoneSerge (back on matrix) | connor (he/him) (UTC-7): 8am instead of 7, rsvp? | 16:01:29 |
connor (burnt/out) (UTC-8) | Yes, 8am Pacific | 17:22:55 |
| @gmacon:matrix.org left the room. | 17:53:23 |
| 15 Oct 2025 |
danielrf | connor (he/him) (UTC-7): Hey, just fyi. This looks very similar to an issue we had fixed in jetpack-nixos: https://github.com/NixOS/nixpkgs/issues/451912 I can't recall if our fix was generic enough to also be applicable to the nixpkgs' nvidia-container-toolkit | 03:56:45 |
connor (burnt/out) (UTC-8) | pain | 07:03:11 |
connor (burnt/out) (UTC-8) | Yeah Jared had written a udevadm settle for some devices; when I refactored to use upstream’s container toolkit stuff I commented it out hoping waiting on the modprobe nvgpu service was enough | 07:05:36 |
connor (burnt/out) (UTC-8) | Relevant PRs:
- https://github.com/anduril/jetpack-nixos/pull/317
- https://github.com/anduril/jetpack-nixos/pull/331
| 07:06:36 |
Gaétan Lepage | connor (he/him) (UTC-7) SomeoneSerge (back on matrix) Torch 2.9.0 was just released.
I'm working on the bump.
It requires libnvshmem_host.so.3. I never heard of OpenSHMEM before.
Do we have it already in nixpkgs? | 18:50:04 |
apyh | oh i have this in a fork, sec | 19:43:52 |
| Ari Lotter joined the room. | 19:44:18 |
Ari Lotter | (still me sorry, diff devices w bad key management) | 19:44:26 |
Ari Lotter | https://github.com/PsycheFoundation/psyche/blob/main/nix/nvshmem.nix | 19:44:30 |
Ari Lotter | we don't have nvshmem in nixpkgs | 19:44:35 |
Ari Lotter | i'm using this with torch 2.9.0 :) | 19:45:38 |
Gaétan Lepage | Thanks for sharing! I guess we'll need to cleanly upstream this then? | 20:03:40 |
Gaétan Lepage | * Thanks for sharing! I guess we'll need to cleanly upstream this then. | 20:03:41 |
connor (burnt/out) (UTC-8) | I could probably add it to the CUDA 13 PR; nvshmem is one of the dependencies of libcublasmp I didn’t try to package | 20:53:02 |
| 16 Oct 2025 |
Ari Lotter | lmk if i can help - 2.9.0 (nightly) is in active usage in the above project | 00:46:36 |
connor (burnt/out) (UTC-8) | Looks like it should be very doable to package — it’s a redist so shouldn’t be too bad and can re-use all the helpers we’ve got for that. Will take a closer look tomorrow | 04:53:50 |
Niclas Overby Ⓝ | Is there something like rust-overlay for CUDA, so you can specify exactly which CUDA version to use? | 11:39:49 |
connor (burnt/out) (UTC-8) | You can specify which CUDA version to use currently so long as it is a CUDA version supported in-tree by using an overlay or the cudaPackages.pkgs pattern (see Nixpkgs manual) | 14:08:11 |
connor (burnt/out) (UTC-8) | Arbitrary versions aren’t something doable with the current state of things because of the amount of patching required for each package (which varies by version of course) | 14:10:31 |
connor (burnt/out) (UTC-8) | Plus, some of these binaries aren’t stand-alone — NVCC for example requires a host compiler.
So if we wanted to support arbitrary CUDA versions, we’d need to somehow know ahead of time which host compilers and versions are supported by all NVCC releases (we have a list we maintain in tree but it’s updated manually by reading release notes).
And then we’d need to use the appropriate version of the host compiler… but what if it’s not packaged in Nixpkgs? CUDA 11 releases used GCC 9, 10, and 11 and those aren’t maintained in-tree any more. | 14:14:59 |
connor (burnt/out) (UTC-8) | I’ve been working on the ability to extend the CUDA package set and make new ones for out of tree users, but it’s generally non-trivial and requires a great deal of familiarity | 14:17:53 |
| 17 Oct 2025 |
connor (burnt/out) (UTC-8) | The CUDA 13 PR now has libnvshmem, built from source (that was not fun)
It does not have nvshmem4py since that’s gonna be irritating to build and requires stuff we don’t have packaged yet | 01:16:43 |