NixOS CUDA | 328 Members | |
| CUDA packages maintenance and support in nixpkgs | https://github.com/orgs/NixOS/projects/27/ | https://nixos.org/manual/nixpkgs/unstable/#cuda | 64 Servers |
| Sender | Message | Time |
|---|---|---|
| 1 Mar 2023 | ||
| and has failing t ests | 03:15:28 | |
| * and has failing tests | 03:15:37 | |
| 03:16:29 | |
| I'ma sleep | 03:16:37 | |
| FRidhSomeone S please tell me if https://github.com/NixOS/nixpkgs/pull/218929 is acceptable | 03:48:17 | |
In reply to @hexa:lossy.network
What the actual fuck, they won't just relax pinned versions? | 12:05:56 | |
In reply to @justbrowsing:matrix.org Word of warning, I haven't been working on this too long but here's what I've noticed. All, please feel free to correct if any of this is wrong.
I believe it's organic -- outside of https://github.com/ryantm/nixpkgs-update I'm not sure what exists in the way of automation for updating nixpkgs.
Given the switch to redistributables and Nix parsing JSON we grab from NVIDIA's website, maybe it'd be easier to automate it now! I could imagine a script which curls their index to see what the latest is, and adds a new copy of the JSON files we need if there's a newer version.
I think we find out about it either from breakages or reading the release notes/JSON. Kevin Mittman I like these questions -- can I add them to the docs tracking issue I have here? https://github.com/NixOS/nixpkgs/issues/217780 | 14:29:48 | |
In reply to @ss:someonex.net With respect to specifying capabilities -- some packages (glares at magma) don't support every capability: https://github.com/NixOS/nixpkgs/pull/217410/files#diff-1e7812b78446dca0e64c4bb933e9255fca6f6539ec1ecd610edf1285a3fcbc56R55 Like, the hell? Skipping 8.6, 8.7, and 8.9? Packages like that make me think we need some way for the coda configuration and derivation to interact to agree on a list of architectures to build for. Imaging setting Or maybe that's desirable? Maybe it would be more annoying if the package used the greatest common factor (say, 8.0 when 8.6 was requested) and no errors were thrown? Is that misleading the user? | 14:44:12 | |
| https://github.com/NixOS/nixpkgs/pull/218265 tf builds ✅ | 15:58:31 | |
In reply to @connorbaker:matrix.orgI was thinking more along the lines of an issue filed here https://github.com/NVIDIA/build-system-archive-import-examples | 16:10:04 | |
In reply to @justbrowsing:matrix.orgIn principle, as long as the directory listing at https://developer.download.nvidia.com/compute/cuda/redist/ works, we could work out our own automation. That being said, a single machine-readable entrypoint (a stable location with a json that lists URIs to all releases, or an RSS feed) would be more convenient | 16:14:00 | |
Been thinking that too. Not even automated PRs, but we could improve visiblity by just making a github workflow that runs on cron schedule, checks published JSONs, and publishes a status report on github pages | 16:16:31 | |
I thought it would be a good idea to run nixpkgs-review with cudaSupport = true, but that just opened a hellgate: https://gist.github.com/SomeoneSerge/6cc00b41964e43f725fc12046778532d#file-218265-log-L23 | 22:38:52 | |
adapta-gtk-theme, ..., chromium, ..., wine64, ... | 22:39:38 | |
| Mostly outPaths react to cudaSupport through gst-plugins-bad, which depend on opencv | 22:40:23 | |
| 😢 | 22:40:55 | |
is that the nix equivalent of portage's emerge world after changing USE flags? | 22:51:03 | |
In reply to @justbrowsing:matrix.orgI'm not familiar with Gentoo, but I think I can guess the answer | 23:44:51 | |
| Tbh, I'd be happy enough to build gst bad against cpu-only opencv, and then just fork the output and patchelf it to use opencv-cuda. Ad hoc as it is, it'd probably discard 90% of false-positives. Just need a good tool for "forking"... | 23:51:45 | |
| 2 Mar 2023 | ||
| Ugh yeah changes which touch OpenCV are such a pain. I flipped when I saw my PR needed me to rebuild 700+ packages | 00:06:44 | |
In reply to @FRidh:matrix.orgFrom a quick look, couldn't tell if this was implemented after all. Do security updates result in massive rebuilds today, or are they shipped using some of these tricks? | 14:53:02 | |
In reply to @connorbaker:matrix.orgOne more thing I'm afraid of: if we go with defaulting to very few capabilities (TM), I imagine the UX of a new user is basically: spend close to an hour on download, find out the stuff you downloaded doesn't work with your graphics | 18:40:23 | |
| Someone on Discourse mention the other day that unpacking the cudatoolkit's runfile took them about 40 minutes (we should get rid of it already) | 18:40:56 | |
| * Someone on Discourse mentioned the other day that unpacking the cudatoolkit's runfile took them about 40 minutes (we should get rid of it already) | 18:41:03 | |
| Download size is another reason for single-capability default, although it's laughable when we use the run-file | 18:45:13 | |
Still. Tensorflow built for 8.6 only is 554.24MiB versus 918.59MiB for "all supported caps" | 18:46:55 | |
| (laughable because this doesn't account for runtime dependencies, which are 9GiB in total) | 18:47:35 | |
* (laughable because this doesn't account for runtime dependencies, which are ~~9GiB in total~~ 9GiB for 8.6-only and 11.15 for [3.5, ..., 8.6]) | 18:48:12 | |
* (laughable because this doesn't account for runtime dependencies, which are ~~9GiB in total~~ 9GiB for 8.6-only and 11.15GiB for [3.5, ..., 8.6]) | 18:48:23 | |
Oh my god, pytorch has migrated to FindCUDAToolkit.cmake!!!!!!!!!!! | 19:09:28 | |