!eWOErHSaiddIbsUNsJ:nixos.org

NixOS CUDA

287 Members
CUDA packages maintenance and support in nixpkgs | https://github.com/orgs/NixOS/projects/27/ | https://nixos.org/manual/nixpkgs/unstable/#cuda58 Servers

Load older messages


SenderMessageTime
4 Mar 2025
@stick:matrix.orgstickan ideal thing for me would be if someone indicated the list of packages that really need to have the build fixed before the merge happens and I would (try to) work on fixing these10:53:23
@stick:matrix.orgstick* an ideal thing for me would be if someone indicated the list of packages that really need to have the build fixed before the merge happens and I will (try to) work on fixing these10:53:33
@ss:someonex.netSomeoneSerge (back on matrix) In addition to Connor's suggestions, can you check what is the output when you run cudaPackages.saxpy? 11:26:55
@stick:matrix.orgstickMaybe the merge of this PR should happen shortly after the merge of ROCm update in #367695 to not do massive rebuilds two times?12:12:00
@stick:matrix.orgstick* Maybe the merge of this PR should happen shortly after the merge of ROCm update in #367695 to not do massive rebuilds two times? https://github.com/NixOS/nixpkgs/pull/36769512:12:17
5 Mar 2025
@angleangleside:matrix.orgasa set a profile picture.08:07:57
@angleangleside:matrix.orgasa changed their display name from Asa to asa.08:08:10
7 Mar 2025
@mdietrich:matrix.orgmdietrich joined the room.13:03:38
@mdietrich:matrix.orgmdietrich Hey all, first of all thank you for your work, last time I tried to use any cuda-related programs and services I had to give up because this joint effort had not been set up.
I am just wondering if I am doing something wrong when trying to set up llama-cpp and open-webui on my NixOS machine. I've set up the nix-community cache (and ollama with CUDA support installs fine in any case), but either enabling nixpkgs.config.cudaSupport or overwriting e.g. llama-cpp's package with `services.llama-cpp.package = pkgs.overwrite { config.cudaSupport = true; config.rocmSupport = false; }
13:12:26
@mdietrich:matrix.orgmdietrich * Hey all, first of all thank you for your work, last time I tried to use any cuda-related programs and services I had to give up because this joint effort had not been set up.
I am just wondering if I am doing something wrong when trying to set up llama-cpp and open-webui on my NixOS machine. I've set up the nix-community cache (and ollama with CUDA support installs fine in any case), but neither enabling nixpkgs.config.cudaSupport or overwriting e.g. llama-cpp's package with `services.llama-cpp.package = pkgs.overwrite { config.cudaSupport = true; config.rocmSupport = false; }` just dowload and install the appropriate packages, but lead to extremely long build times. Are these packages (llama-cpp and open-webui, of which I think onnxruntime takes the longest) just not built in the community cache?
13:13:54
@ss:someonex.netSomeoneSerge (back on matrix)Let's see13:15:53
@ss:someonex.netSomeoneSerge (back on matrix)https://hydra.nix-community.org/job/nixpkgs/cuda/llama-cpp.x86_64-linux13:15:55
@ss:someonex.netSomeoneSerge (back on matrix)https://hydra.nix-community.org/job/nixpkgs/cuda/onnxruntime.x86_64-linux13:16:26
@ss:someonex.netSomeoneSerge (back on matrix) open-webui apparently wasn't added to the release-cuda.nix file yet: https://hydra.nix-community.org/job/nixpkgs/cuda/open-webui.x86_64-linuxd 13:17:10
@ss:someonex.netSomeoneSerge (back on matrix)As for onnxruntime and llama-cpp, let's compare the hashes in your llama-cpp and the one reported by hydra13:18:20
@mdietrich:matrix.orgmdietrich

I am on x86_64, nixos-unstable with flakes with an RTX 3060 Ti and following substituters:

        substituters = [
          "https://nix-community.cachix.org"
        ];
        trusted-public-keys = [
          "nix-community.cachix.org-1:mB9FSh9qf2dCimDSUo8Zy7bkq5CX+/rkCWyvRCYg3Fs="
        ];
13:19:29
@mdietrich:matrix.orgmdietrichThank you for your quick answer13:20:01
@mdietrich:matrix.orgmdietrichI think onnxruntime is a dependency of open-webui and not llama-cpp, open-webui itself probably (?) does not need cuda support itself13:20:45
@mdietrich:matrix.orgmdietrich services.llama-cpp.package has the value "«derivation /nix/store/dhqdwqp6akr6h6f1k3rz190m3syrv6iy-llama-cpp-4731.drv»" 13:23:06
@ss:someonex.netSomeoneSerge (back on matrix) Let's try nix path-info --override-input nixpkgs github:NixOS/nixpkgs/1d2fe0135f360c970aee1d57a53f816f3c9bddae --derivation .#nixosConfigurations.$(hostname).config.services.llama-cpp.package to make it comparable with https://hydra.nix-community.org/build/3552955#tabs-buildinputs 13:24:11
@ss:someonex.netSomeoneSerge (back on matrix)I'd maybe not focus on these concerns, the expert hours are arguably more expensive that rebuild costs13:25:27
@ss:someonex.netSomeoneSerge (back on matrix) (still pending =) 13:25:51
@mdietrich:matrix.orgmdietrich Wait a minute, I am slightly confused as llama-cpp seems to actually have cuda support now that I rebuilt a couple of minutes ago. It just does not use my GPU when running inference even though it reports it as visible and usable. Maybe a configuration mistake on my side (although I am using the default NixOS service). I'll look into open-webui and onnxruntime now... 13:27:31
@mdietrich:matrix.orgmdietrichYes, onnxruntime does recompile, as well as python3.12-torch-2.5.1. I'm checkin the hashes now...13:33:36
@mdietrich:matrix.orgmdietrich

I am definitely building onnxruntime myself even though I get:

> nix path-info --override-input nixpkgs github:NixOS/nixpkgs/9f41a78ead0fbe2197cd4c09b5628060456cd6e3 --derivation .\#nixosConfigurations.$(hostname).pkgs.onnxruntime
• Updated input 'nixpkgs':
    'github:nixos/nixpkgs/32fb99ba93fea2798be0e997ea331dd78167f814?narHash=sha256-ozoOtE2hGsqh4XkTJFsrTkNxkRgShxpQxDynaPZUGxk%3D' (2025-02-21)
  → 'github:NixOS/nixpkgs/9f41a78ead0fbe2197cd4c09b5628060456cd6e3?narHash=sha256-WWXRCTOWcKvtzqzVgBMON0/TWcFMyWq831HQUITE4rs%3D' (2025-02-21)
/nix/store/a22vqi9d0ndhlcy1yxw4m3ir4z7ckfrg-onnxruntime-1.20.1.drv

Which is the same hash as the hydra build store path

13:48:48
@mdietrich:matrix.orgmdietrichI get the same hash for pytorch locally and in hydra as well!13:56:11
@ss:someonex.netSomeoneSerge (back on matrix) And if you nix build --override-input nixpkgs github:NixOS/nixpkgs/9f41a78ead0fbe2197cd4c09b5628060456cd6e3 .#nixosConfigurations.$(hostname).pkgs.onnxruntime? 13:59:42
@mdietrich:matrix.orgmdietrichThen I'm building nccl and cudnn-frontend for some reason?14:15:15
@ss:someonex.netSomeoneSerge (back on matrix)Well this certainly shouldn't be happening if the hashes indeed match14:21:40
@ss:someonex.netSomeoneSerge (back on matrix)Which hydra eval did you refer to?14:22:00

Show newer messages


Back to Room ListRoom Version: 9