!eWOErHSaiddIbsUNsJ:nixos.org

NixOS CUDA

289 Members
CUDA packages maintenance and support in nixpkgs | https://github.com/orgs/NixOS/projects/27/ | https://nixos.org/manual/nixpkgs/unstable/#cuda57 Servers

Load older messages


SenderMessageTime
14 Nov 2024
@ss:someonex.netSomeoneSerge (back on matrix) * i haven't even had time to review zeuner's tensorflow prs...16:46:03
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8)
In reply to @aliarokapis:matrix.org
hi all! is https://docs.nvidia.com/vpi/2.0/index.html packaged anywhere?
as far as I can tell, it's only available via debian installers. I'd take a look at https://github.com/anduril/jetpack-nixos since they have some tooling set up to repackage stuff like that (they also already have VPI iirc, but maybe not a new version)
17:40:58
15 Nov 2024
@aliarokapis:matrix.orgAlexandros LiarokapisI wish I could run nixos modules inside a light container or something on as non-nixos machine. Waiting for systemd-nspawn to make this easier.09:46:58
@adam:robins.wtf@adam:robins.wtf What’s a light container?  12:46:54
@aliarokapis:matrix.orgAlexandros LiarokapisNothing I mean just a container, compared to a full blown vm.12:49:52
@hexa:lossy.networkhexacool, we'll also break numba18:48:28
@adam:robins.wtf@adam:robins.wtf Alexandros Liarokapis: i do this with incus. i'd probably call it a fat container, but it's a full nixos env inside a container. the downside/upside is it's separately managed from the host 22:00:55
16 Nov 2024
@aliarokapis:matrix.orgAlexandros LiarokapisThis is very interesting. I recall a redhat talk modifying LXD to achieve this and that is shrn I started looking into it08:05:21
@aliarokapis:matrix.orgAlexandros LiarokapisGot any respurces I can look into?08:05:30
@aliarokapis:matrix.orgAlexandros Liarokapis* This is very interesting. I recall a redhat talk modifying LXD to achieve this and that is when I started looking into it08:05:38
@aliarokapis:matrix.orgAlexandros Liarokapis* Got any resources I can look into?08:06:15
@aliarokapis:matrix.orgAlexandros LiarokapisActially I think the wiki page has enough info to get me started08:06:31
@aliarokapis:matrix.orgAlexandros Liarokapis* Actually I think the wiki page has enough info to get me started08:06:41
@aliarokapis:matrix.orgAlexandros Liarokapis.. or not, it is mainly nixos based.08:15:50
@aliarokapis:matrix.orgAlexandros Liarokapisi guess I may as well try it08:16:07
@hexa:lossy.networkhexa
   error: tensorflow-gpu-2.13.0 not supported for interpreter python3.12
20:45:57
@hexa:lossy.networkhexathe sound of nixos 24.05 hits hard20:46:03
@hexa:lossy.networkhexa *
   error: tensorflow-gpu-2.13.0 not supported for interpreter python3.12
20:46:08
@hexa:lossy.networkhexa *
error: tensorflow-gpu-2.13.0 not supported for interpreter python3.12
20:46:12
17 Nov 2024
@glepage:matrix.orgGaétan LepageYes... Let's hope zeuner finds the time to end the TF bump...10:38:39
18 Nov 2024
@hexa:lossy.networkhexa
 wyoming-faster-whisper[4505]:   File "/nix/store/dfp38l0dy3n97wvrgz5i62mwvsmshd3n-python3.12-faster-whisper-unstable-2024-07-26/lib/python3.12/site-packages/faster_whisper/transcribe.py", line 145, in __init__
 wyoming-faster-whisper[4505]:     self.model = ctranslate2.models.Whisper(
 wyoming-faster-whisper[4505]:                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^
 wyoming-faster-whisper[4505]: RuntimeError: CUDA failed with error unknown error
 systemd[1]: wyoming-faster-whisper-medium-en.service: Main process exited, code=exited, status=1/FAILURE
02:09:21
@hexa:lossy.networkhexaalso loving unknown error errors02:09:26
@hexa:lossy.networkhexa
wyoming-faster-whisper[4745]:   File "/nix/store/dfp38l0dy3n97wvrgz5i62mwvsmshd3n-python3.12-faster-whisper-unstable-2024-07-26/lib/python3.12/site-packages/faster_whisper/transcribe.py", line 145, in __init__
wyoming-faster-whisper[4745]:     self.model = ctranslate2.models.Whisper(
wyoming-faster-whisper[4745]:                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^
wyoming-faster-whisper[4745]: RuntimeError: CUDA failed with error no CUDA-capable device is detected
02:10:44
@hexa:lossy.networkhexababy steps02:10:46
@hexa:lossy.networkhexaI can confirm the card is still seated correctly 😄 02:10:58
@hexa:lossy.networkhexahardening at work02:18:46
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8)Ugh I don’t like computers05:10:46
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8)

Anyway in the interest of splitting my attention ever more thinly I decided to start trying to work on some approach toward evaluation of derivations and building them
The idea being to have

  1. a service which is given a flake ref and an attribute path and efficiently produces a list of attribute paths to derivations exiting under the given attribute path and stores the eval time somewhere
  2. a service which is given a flake ref and an attribute path to a derivation and produces the JSON representation of the closure of derivations required to realize the derivation, again storing eval time somewhere
  3. a service which functions as a job scheduler, using historical data about costs (space, time, memory, CPU usage, etc.) and information about locality (existing store paths on different builders) to realize a derivation, which is updated upon realization of a derivation
05:18:41
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8)Because why have one project when you can have many?05:18:55
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8)

https://github.com/ConnorBaker/nix-eval-graph

And I’ve decided to write it in Rust, which I am self teaching.

And I’ll probably use a graph database, because why not.

And I’ll use NixOS tests for integration testing, because also why not.

05:20:02

Show newer messages


Back to Room ListRoom Version: 9