| 4 Nov 2025 |
Ari Lotter | i'm so tired of CUDA nightmares 😠im so close to giving up and building dockerized devenvs, i just really don't want to give in..... :( | 22:37:57 |
Gaétan Lepage | (It's a secret, but you might want to add https://cache.nixos-cuda.org as a substituter, it is slowly getting more and more artifacts)
Public key: cache.nixos-cuda.org:74DUi4Ye579gUqzH4ziL9IyiJBlDpMRn9MBN8oNan9M= | 22:44:02 |
Gaétan Lepage | connor (burnt/out) (UTC-8), Serge and I got #457803 ready.
We are waiting for nixpkgs's CI to get fixed (https://github.com/NixOS/nixpkgs/pull/458647).
Let's merge ASAP | 23:38:07 |
Robbie Buxton | For flash attention you should use the version of cutlass in the repo | 23:54:57 |
Robbie Buxton | They have a hash | 23:55:06 |
Robbie Buxton | In csrc/cutlass | 23:56:01 |
Robbie Buxton | * They have a rev | 23:56:25 |
| 5 Nov 2025 |
apyh | ah fair enough | 00:10:30 |
SomeoneSerge (matrix works sometimes) | step 1: torchWithCuda = pkgsCuda.....torch (we were supposed to be here now, but it got out of hand) step 2: torchWithCuda = warn "..." pkgsCuda... step 3: torchWithCuda = throw | 00:12:18 |
SomeoneSerge (matrix works sometimes) | and what we really want is late binding and incremental builds | 00:13:41 |
connor (burnt/out) (UTC-8) | Why are you building for so many CUDA capabilities? I can’t really think of a reason you’d need that range in particular. | 01:59:14 |
connor (burnt/out) (UTC-8) | Added to merge queue | 02:07:23 |
apyh | In reply to @connorbaker:matrix.org Why are you building for so many CUDA capabilities? I can’t really think of a reason you’d need that range in particular. 's a distributed ml training application that needs to run on everything from gtx 10xx gpus to modern data center GH/GB200s :/ | 03:27:37 |
apyh | most common hardware is gonna be 30xx 40xx 50xx, h100, a100, b200 | 03:27:56 |
apyh | though.. i could just see what pytorch precompiled wherls runs on and limit to that | 03:28:54 |
apyh | should be fine | 03:28:56 |
connor (burnt/out) (UTC-8) | Flash attention doesn’t support anything older than Ampere I thought | 03:29:07 |
Robbie Buxton | V2 does | 03:29:19 |
Robbie Buxton | V3 is hopper only | 03:29:24 |