!eWOErHSaiddIbsUNsJ:nixos.org

NixOS CUDA

279 Members
CUDA packages maintenance and support in nixpkgs | https://github.com/orgs/NixOS/projects/27/ | https://nixos.org/manual/nixpkgs/unstable/#cuda57 Servers

Load older messages


SenderMessageTime
12 Nov 2025
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8) SomeoneSerge (back on matrix)Gaétan LepageI added you both as collaborators on my fork of Nixpkgs. I had a death in the family so I’m going to be in and out. We need to get this PR fixed up and merged: https://github.com/NixOS/nixpkgs/pull/456510
We also need to get this PR merged: https://github.com/NixOS/nixpkgs/pull/459416
I likely won’t have time to do that over the next few days. Can you two do that for me?
19:12:06
@leona:leona.isleonatake care <320:40:59
@daniel-fahey:matrix.orgDaniel FaheyCondolences21:24:01
@arilotter:matrix.orgAri Lotteryeah, givin up - jax dies every time :(21:34:25
@daniel-fahey:matrix.orgDaniel Faheyhaha, well done for trying22:19:37
@glepage:matrix.orgGaétan Lepage I'm so sorry Connor. Take your time. Focus on what matters the most for you right now (i.e. not CUDA).
I'll look at those two PRs. Please message me if I can do anything else.
23:04:38
@arilotter:matrix.orgAri Lotteryeaaah, this is somethin' real broken with the torch 2.9.0 update :/ gonna see if i can figure it out, but it just doesn't seem that nvrtc is in the run path. looking fora way to repro it without my whole binary 😭23:17:18
@glepage:matrix.orgGaétan LepageRedacted or Malformed Event23:29:07
@glepage:matrix.orgGaétan Lepage torch 2.9.1 is out. And triton 3.5.1. And both staging-next and staging-nixos were merged into master a few hours ago.
CPUs will have to work hard for the next few days...
23:29:21
13 Nov 2025
@ss:someonex.netSomeoneSerge (back on matrix)They come not single spies... really sorry to hear this, Connor. Take care14:07:48
@arilotter:matrix.orgAri Lotteroh woof, but torch-bin is 2.9.1 and torch is still 2.9.015:10:16
@arilotter:matrix.orgAri Lotteraha, repro'd :D16:15:41
@arilotter:matrix.orgAri Lotterhttps://github.com/NixOS/nixpkgs/issues/461334 issue opened :)18:54:11
14 Nov 2025
@hexa:lossy.networkhexa (UTC+1)https://hydra.nixos-cuda.org/build/14219 magma runs into the output limit04:50:01
@hexa:lossy.networkhexa (UTC+1)and https://hydra.nixos-cuda.org/jobset/nixos-cuda/cuda-packages-v2#tabs-jobs has no torch package 🤔04:50:51
@glepage:matrix.orgGaétan LepageI increased it from 4GB (what nix-community has I think) to 8GB. And it seems to still be broken...08:53:41
@glepage:matrix.orgGaétan LepageThis is very weird. It ends up being built anyway as a dependency. I'll try to investigate...08:55:38
@glepage:matrix.orgGaétan Lepage Ok, I figured it out. torch and torchWithoutRocm have the same outPaths. So torch is getting filtered out in favor of torchWithoutRocm. 09:25:13
@arilotter:matrix.orgAri Lotter realized this isn't a 2.9 regression, it's a -bin vs source problem :/ 18:37:14
@arilotter:matrix.orgAri Lotterbin works fine T_T18:37:19
@arilotter:matrix.orgAri Lotterupdated the ticket :)18:37:29
@glepage:matrix.orgGaétan Lepage I updated torch-bin to 2.9.1 yesterday. The PR for the source-based build is https://github.com/NixOS/nixpkgs/pull/461241 21:55:22
@apyh:matrix.orgapyhi see your commit message says torch 2.8->2.9, but it's actually 2.9->2.9.1 :)21:56:32
@glepage:matrix.orgGaétan LepageGood catch, now fixed.22:06:15
15 Nov 2025
@cafkafk:gitter.imcafkafk joined the room.12:47:57
@glepage:matrix.orgGaétan Lepage SomeoneSerge (back on matrix) would you have a minute to take a look at the triton/torch bump?
https://github.com/NixOS/nixpkgs/pull/461241
14:23:53
@glepage:matrix.orgGaétan LepageBuilt with and without CUDA. No obvious regressions.14:24:19
17 Nov 2025
@bjth:matrix.orgBryan Honof

How would you go about conditionally setting cudaCapabilities when instantiating nixpkgs? I.e.

Image I have this.

{
  inputs = {
    nixpkgs = "github:nixos/nixpkgs?ref=nixos-25.05";
  };
  
  outputs = { self, nixpkgs }: {
    packages.x86_64-linux.default = let
      pkgs = import nixpkgs {
        overlays = [ ];
        config = {
          allowUnfree = true;
          cudaSupport = true;
          cudaCapabilities = [ "..." "..." ];
        };
      };
    in
    pkgs.hello;
    
    packages.aarch64-linux.default = let
      pkgs = import nixpkgs {
        overlays = [ ];
        config = {
          allowUnfree = true;
          cudaSupport = true;
          cudaCapabilities = if isJetson then [ "..." "..." ] else [ "..." "..." ];
        };
      };
    in
    pkgs.hello;
  };
}

It's the aarch64-linux part specifically that I'm a bit stuck on. I have some cloud servers that have an NVIDIA GPUs in them that run aarch64-linux, but I also have some Jetson devices that are also considered aarch64-linux.

And if I understand the whole thing correctly, I can't just set the cudaCapabilities list to include both the non-jetson and jetson capabilities, right? Or at least, than isJetsonBuild would just always eval to true even if the build was meant for the cloud server.

Probably something stupid I'm just overlooking, sorry for bothering. 😅

17:35:32
@ss:someonex.netSomeoneSerge (back on matrix)

It's the aarch64-linux part specifically that I'm a bit stuck

There's aarch64-linux and there's aarch64-linux. It's an artifact of us not including cuda/rocm stuff in hostPlatform (yet). The isJetsonBuild should only evaluate to true if your cudaCapabilities are jetson capabilities

19:43:44
@ss:someonex.netSomeoneSerge (back on matrix) So it's not really about "setting cudaCapabilities conditionally", it's about instantiating nixpkgs for different platforms. For flakes you'd have to suffix the attributes of one of the aarch64-linux platforms, or move stuff to legacyPackages, but, of course, you could also simply not maintain the list of already-evaluated and not-really-overridable "recipes", i.e. drop the flake:) 19:47:42

Show newer messages


Back to Room ListRoom Version: 9