!eWOErHSaiddIbsUNsJ:nixos.org

NixOS CUDA

312 Members
CUDA packages maintenance and support in nixpkgs | https://github.com/orgs/NixOS/projects/27/ | https://nixos.org/manual/nixpkgs/unstable/#cuda61 Servers

Load older messages


SenderMessageTime
9 Jun 2024
@connorbaker:matrix.orgconnor (he/him)It works without compile, just curious if this is a problem with the PR20:50:43
@connorbaker:matrix.orgconnor (he/him)I'll try again with master to make sure it's not a regression20:51:08
@connorbaker:matrix.orgconnor (he/him)Cool, fails on master too20:51:57
@glepage:matrix.orgGaétan LepageNice ^^20:53:19
@glepage:matrix.orgGaétan LepageIs it OK for me to merge now ?20:53:44
@glepage:matrix.orgGaétan LepageOh, I've just seen your message20:53:58
10 Jun 2024
@mjolnir:nixos.orgNixOS Moderation Bot unbanned @jonringer:matrix.org@jonringer:matrix.org.00:17:14
@glepage:matrix.orgGaétan Lepageclipboard.png
Download clipboard.png
06:44:40
@glepage:matrix.orgGaétan LepageHaha botorch has probably taken ~11h but it succeeded X)06:44:56
@shekhinah:she.khinah.xyzshekhinah set their display name to yaldebaoth.11:02:59
@shekhinah:she.khinah.xyzshekhinah changed their display name from yaldebaoth to yaldabaoth.11:03:43
@connorbaker:matrix.orgconnor (he/him) Gaétan Lepage: did you mention there was a PR or something merged to disable the checkPhase or test suite for botorch, or did I misunderstand? 14:01:56
@connorbaker:matrix.orgconnor (he/him) On another note, has anyone built elpa (https://github.com/NixOS/nixpkgs/blob/master/pkgs/development/libraries/elpa/default.nix) successfully with CUDA support? I let it run for like 20h and it was still building. Seems to compile four object files at a time? 14:04:03
@glepage:matrix.orgGaétan Lepage
In reply to @connorbaker:matrix.org
Gaétan Lepage: did you mention there was a PR or something merged to disable the checkPhase or test suite for botorch, or did I misunderstand?
No, I have not done anything. I was actually able to build it just fine from master earlier today.
14:29:01
@hexa:lossy.networkhexa Gaétan Lepage: have you considered pulling this patch for tensorflow-bin? https://github.com/tensorflow/tensorflow/issues/58073#issuecomment-2097055553 20:58:34
11 Jun 2024
@keiichi:matrix.orgteto when using localai 2.15 from unstable and even after a reboot I get ggml_cuda_init: failed to initialize CUDA: CUDA driver is a stub library. It's a bit random but if anyone has a tip, I take it. nvidia-smi output looks fine 00:25:38
@glepage:matrix.orgGaétan Lepage
In reply to @hexa:lossy.network
Gaétan Lepage: have you considered pulling this patch for tensorflow-bin? https://github.com/tensorflow/tensorflow/issues/58073#issuecomment-2097055553
This looks like it could work !
However, how do you apply a patch to a wheel-type python derivation ?
06:38:47
@glepage:matrix.orgGaétan Lepage What phase of the buildPythonPackage script should I hook it into ? 06:39:02
@glepage:matrix.orgGaétan Lepage I tried patches = [ but it does not work 06:39:15
@glepage:matrix.orgGaétan Lepage

I am packaging this: https://github.com/EricLBuehler/mistral.rs?tab=readme-ov-file#installation-and-build
You can see that it support several variations for building (CUDA, metal, mkl...)

-> What should be the approach ? Adding cudaSupport ? metalSupport ? mklSupport ?

07:01:41
@kaya:catnip.eekaya 𖤐 changed their profile picture.08:03:48
@hexa:lossy.networkhexa
In reply to @glepage:matrix.org
This looks like it could work !
However, how do you apply a patch to a wheel-type python derivation ?
likely in postInstall 😕
11:58:18
@hexa:lossy.networkhexacurses11:59:30
@glepage:matrix.orgGaétan Lepage Ok, but I can I use fetchpatch though ? 12:02:43
@ss:someonex.netSomeoneSerge (matrix works sometimes) connor (he/him) (UTC-5) IIRC you brought up setting legacy (FindCUDA&c) variables from the setup hooks. I think we should set them, and we should put that logic behind a guard (e.g. findCudaCmakeSupport=true), just as we should guard the current logic (e.g. findCudatoolkitCmakeSupport=true). We should disable the legacy by default. We should only set cmake flags when the cmake hook is actually used or when cmake flags are explicitly requested. 13:19:22
@ss:someonex.netSomeoneSerge (matrix works sometimes)
In reply to @keiichi:matrix.org
when using localai 2.15 from unstable and even after a reboot I get ggml_cuda_init: failed to initialize CUDA: CUDA driver is a stub library. It's a bit random but if anyone has a tip, I take it. nvidia-smi output looks fine
LD_DEBUG=libs
13:19:44
@ss:someonex.netSomeoneSerge (matrix works sometimes)
In reply to @glepage:matrix.org

I am packaging this: https://github.com/EricLBuehler/mistral.rs?tab=readme-ov-file#installation-and-build
You can see that it support several variations for building (CUDA, metal, mkl...)

-> What should be the approach ? Adding cudaSupport ? metalSupport ? mklSupport ?

Does it allow enabling multiple features at once?
13:20:13
@glepage:matrix.orgGaétan Lepage
In reply to @ss:someonex.net
Does it allow enabling multiple features at once?
No, but I think that I will copy the implementation from ollama
13:20:38
@glepage:matrix.orgGaétan LepageIt looks very clean to me13:20:44
@glepage:matrix.orgGaétan Lepage https://github.com/NixOS/nixpkgs/blob/master/pkgs/by-name/ol/ollama/package.nix#L65-L82 13:21:07

Show newer messages


Back to Room ListRoom Version: 9