!eWOErHSaiddIbsUNsJ:nixos.org

NixOS CUDA

311 Members
CUDA packages maintenance and support in nixpkgs | https://github.com/orgs/NixOS/projects/27/ | https://nixos.org/manual/nixpkgs/unstable/#cuda61 Servers

Load older messages


SenderMessageTime
16 Apr 2026
@sporeray:matrix.orgRobbie Buxton I was using nixbuild.net and it kept sizing it up and giving it more cores, and yeah the solution was reducing cores obviously 13:38:02
@hexa:lossy.networkhexathat's not friendly at all, but does fit on my desktop still13:38:12
@hexa:lossy.networkhexabut that means your core/ram balance was way off13:38:28
@sporeray:matrix.orgRobbie Buxton
In reply to @hexa:lossy.network
but that means your core/ram balance was way off
How so?
13:39:20
@hexa:lossy.networkhexayou have a base need of RAM per core13:39:51
@hexa:lossy.networkhexajust to run the build on it13:39:58
@hexa:lossy.networkhexaand if you oom the total cores probably needed more ram than what was provided13:40:17
@hexa:lossy.networkhexaand the solution would have been to get fewer cores with more ram instead13:40:30
@sporeray:matrix.orgRobbie BuxtonIdk I can’t get more than 1.6 TB of ram on AWS 😁13:40:42
@hexa:lossy.networkhexanow I don't know anything about how nixbuild.net does scaling13:40:56
@sporeray:matrix.orgRobbie Buxton The large instances have interesting ratios 13:40:55
@hexa:lossy.networkhexayeah, but if you get 10k cores at the same time, then what?13:41:09
@sporeray:matrix.orgRobbie Buxton Iirc it’s not that smart for sizing (the service itself is fab tho) 13:41:32
@hexa:lossy.networkhexathat's 160 MB per core if I didn't miscalculate13:41:38
@sporeray:matrix.orgRobbie Buxton
In reply to @hexa:lossy.network
yeah, but if you get 10k cores at the same time, then what?
On a single machine I can dream
13:41:47
@hexa:lossy.networkhexaright13:41:55
@glepage:matrix.orgGaétan Lepage Yes, our cuda-packages jobset is broken because python3Packages.deep-ep emits a warning at eval time. 13:56:06
@glepage:matrix.orgGaétan Lepage If someone may review https://github.com/NixOS/nixpkgs/pull/510375, we could start building again 13:56:23
@glepage:matrix.orgGaétan LepageSorry everyone for the disruption13:56:35
@ss:someonex.netSomeoneSerge (matrix works sometimes)A reminder that the cache is not an officially maintained cuda team thing at the moment, and only exposed as an accidental dev artifact. Reminder that "stable channels" are not an official cuda team thing. Only hydra itself is. For the time being.14:00:03
17 Apr 2026
@glepage:matrix.orgGaétan Lepage

SomeoneSerge (matrix works sometimes) do you think we can get this in today?
https://github.com/NixOS/nixpkgs/pull/498612#issuecomment-4266520526

It's be nice to have it as part as 26.05 and the window closes by tomorrow

09:31:18
@ss:someonex.netSomeoneSerge (matrix works sometimes)These changes are backportable12:59:28
@ss:someonex.netSomeoneSerge (matrix works sometimes) * 12:59:47
@glepage:matrix.orgGaétan LepageSure, but it'd be easier to merge them now (if they're ready ofc)c13:11:32
@glepage:matrix.orgGaétan Lepage* Sure, but it'd be easier to merge them now (if they're ready ofc)13:11:34
18 Apr 2026
@ccicnce113424:matrix.orgccicnce113424Actually, the window doesn't close until 2026-04-27 07:00 UTC, so we've still got a few days, not "tomorrow". That said, I think this PR is ready to merge right now.09:03:52
20 Apr 2026
@edwtjo:fairydust.spaceEdward TjörnhammarYou risk breaking HPC deployments with these changes09:05:45
@edwtjo:fairydust.spaceEdward Tjörnhammar* You risk breaking HPC deployments with these changes, rushing them without further confirmations is reckless09:07:42
@edwtjo:fairydust.spaceEdward Tjörnhammar* You risk breaking HPC deployments with these changes, rushing them without further confirmations is reckless09:08:01
@ccicnce113424:matrix.orgccicnce113424 @[Edward Tjörnhammar] I'm not rushing anything. I didn't touch the HPC-specific logic, and the PR is literally open for review. I've already replied to all your concerns in the PR comments. If you genuinely believe HPC is broken by these changes, please quote the exact lines of code on GitHub that cause the issue. Let's keep the discussion technical and factual. 09:40:13

Show newer messages


Back to Room ListRoom Version: 9