| 11 Mar 2026 |
Kevin Mittman (jetlagged/UTC-7) | Redacted or Malformed Event | 01:54:11 |
Gaétan Lepage | connor (burnt/out) (UTC-8) would you agree with a 12.8 -> 12.9 global bump before messing around with 13.0? | 11:05:21 |
| Theuni changed their display name from Theuni to Christian Theune. | 14:13:00 |
connor (he/him) | Sure! I remember some weird breakages a while back when I had wanted to bump immediately after 12.9 became available, but hopefully they’re all resolved by now :) | 16:08:54 |
Gaétan Lepage | https://github.com/NixOS/nixpkgs/pull/498861 | 16:43:46 |
Gaétan Lepage | connor (burnt/out) (UTC-8)
About https://github.com/NixOS/nixpkgs/pull/498681, I plan to build torch and vllm. If this works fine, I will merge it.
With the CUDA PRs on the way, I won't have the capacity to exhaustively test all of them.
No objection on your side? | 23:37:24 |
Gaétan Lepage | (same reasonning for https://github.com/NixOS/nixpkgs/pull/498678#issuecomment-4035473707). | 23:39:46 |
Gaétan Lepage | * (same reasonning for https://github.com/NixOS/nixpkgs/pull/498678). | 23:39:52 |
connor (he/him) | Sounds good! I’ll leave a comment on them | 23:51:04 |
Gaétan Lepage | I'm testing the CUDA bump more thoroughly though.
~1.3k rebuilds left (out of 1.8k) | 23:53:19 |
Gaétan Lepage | * connor (burnt/out) (UTC-8)
About https://github.com/NixOS/nixpkgs/pull/498681, I plan to build torch and vllm. If this works fine, I will merge it.
With all the CUDA PRs in the queue, I won't have the capacity to exhaustively test all of them.
No objection on your side? | 23:54:04 |
| 12 Mar 2026 |
| Theuni changed their display name from Christian Theune to Theuni. | 07:18:55 |
Bryan Honof | It looks like torch's supportedTorchCudaCapabilities was out-of-sync with upstream. https://github.com/NixOS/nixpkgs/pull/499216
How would I use nixpkgs-review to test these changes?
| 10:53:19 |
Gaétan Lepage | Thanks for the PR!
Well, you don't want to rebuild all torch consumers for this. What you can do is the following:
nixpkgs-review --extra-nixpkgs-config "{ allowUnfree = true; cudaSupport = true; }" -p python3Packages.torch -p python3Packages.vllm -p python3Packages.torchvision
| 12:39:42 |
Gaétan Lepage | I'll try to have a look at it before next week | 12:39:58 |
Gaétan Lepage | connor (burnt/out) (UTC-8) actually, neither the current nor the new gpu-burn version work:
❮ ./result/bin/gpu_burn
Run length not specified in the command line. Using compare file: /nix/store/9c2avfi2bxc2aydfl2sdgkp8iamhj8as-gpu-burn-0-unstable-2024-04-09/share/compare.ptx
Burning for 10 seconds.
GPU 0: NVIDIA GeForce RTX 3060 (UUID: GPU-7d08a1e6-4634-499f-d58a-91bf77137f69)
Initialized device 0 with 11911 MB of memory (11788 MB available, using 10609 MB of it), using FLOATS
Results are 268435456 bytes each, thus performing 39 iterations
Couldn't init a GPU test: Error in load module (gpu_burn-drv.cpp:239): a PTX JIT compilation failed
0.0% proc'd: -1 (0 Gflop/s) errors: 0 (DIED!) temps: 36 C
(tested on 2 GPUs) | 19:56:18 |
Gaétan Lepage | Nevermind, all good.
You need to carefully set cudaCapabilities for it to run fine on a given GPU: https://github.com/NixOS/nixpkgs/pull/499323#issuecomment-4049769046 | 20:25:50 |
apyh | In reply to @glepage:matrix.org Nevermind, all good.
You need to carefully set cudaCapabilities for it to run fine on a given GPU: https://github.com/NixOS/nixpkgs/pull/499323#issuecomment-4049769046 should it have an isBroken if cudaCapabilities has more than one item, then? | 20:42:25 |
Gaétan Lepage | Not really. It selects the highest (techically, the last) capability from your config.cudaCapabilities.
So there's no fundamental reason why a list with additional, lower caps than the one of your GPU could not work. | 22:14:15 |
| 13 Mar 2026 |
Bryan Honof | Thanks! That seems to have built succesfully. :) | 09:26:19 |
| 16 Mar 2026 |
| Kevin Mittman (jetlagged/UTC-7) changed their display name from Kevin Mittman (jetlagged/UTC+8) to Kevin Mittman (jetlagged/UTC-7). | 00:57:26 |
| 4 Aug 2022 |
| Winter (she/her) joined the room. | 03:26:42 |
Winter (she/her) | (hi, just came here to read + respond to this.) | 03:28:52 |
tpw_rules | hey. i had previously sympathzied with samuela and like i said before had some of the same frustrations. i just edited my github comment to add "[CUDA] packages are universally complicated, fragile to package, and critical to daily operations. Nix being able to manage them is unbelievably helpful to those of us who work with them regularly, even if support is downgraded to only having an expectation of function on stable branches." | 03:29:14 |
Winter (she/her) | In reply to @tpw_rules:matrix.org i'm mildly peeved about a recent merging of something i maintain where i'm pretty sure the merger does not own the expensive hardware required to properly test the package. i don't think it broke anything but i was given precisely 45 minutes to see the notification before somebody merged it ugh, 45 minutes? that's... not great. not to air dirty laundry but did you do what samuela did in the wandb PR and at least say that that wasn't a great thing to do? (not sure how else to word that, you get what i mean) | 03:30:23 |
tpw_rules | no, i haven't yet, but i probably will | 03:31:03 |
Winter (she/her) | i admittedly did that with a PR once, i forget how long the maintainer was requested for but i merged it because multiple people reported it fixed the issue. the maintainer said "hey, don't do that" after and now i do think twice before merging. so it could help, is what i'm saying. | 03:31:50 |
tpw_rules | i'm not sure what went wrong with the wandb PR anyway, i think it was just a boneheaded move on the maintainer's part | 03:32:10 |
Winter (she/her) | (it was also simple enough that it was fine and the maintainer said it looked good after) | 03:32:15 |
tpw_rules | * i'm not sure what went wrong with the wandb PR anyway, i think it was just a boneheaded move on the merger's part | 03:32:19 |