!eWOErHSaiddIbsUNsJ:nixos.org

NixOS CUDA

282 Members
CUDA packages maintenance and support in nixpkgs | https://github.com/orgs/NixOS/projects/27/ | https://nixos.org/manual/nixpkgs/unstable/#cuda58 Servers

Load older messages


SenderMessageTime
8 Mar 2023
@connorbaker:matrix.orgconnor (he/him)Direct link to the project: https://github.com/orgs/NixOS/projects/27 It's a private project because that seems to be the default; I gave the CUDA maintainers team admin access so it could be changed to public if that's of interest. I spent a little bit of time going through open issues on Nixpkgs which are at all related to CUDA and added them to the swim lane named "New." If any don't belong on this project (for example, because it falls outside the scope of this maintainer group), please feel free to remove it. I haven't figured out a way to have new issues tagged with the CUDA topic automatically added to the project, but I figure this is a good start to get some visibility into what's currently open.00:57:41
@ss:someonex.netSomeoneSerge (back on matrix) connor (he/him): that's so great! 00:59:52
@ss:someonex.netSomeoneSerge (back on matrix)Yes, can see it00:59:56
@ss:someonex.netSomeoneSerge (back on matrix)https://github.blog/changelog/2023-01-19-github-issues-january-19th-update/ Seems that we need this special "auto-add to project" workflow01:04:50
@connorbaker:matrix.orgconnor (he/him)Hopefully the "by assignee" and "by status" views will be helpful too (tabular instead of Kanban)01:05:19
@connorbaker:matrix.orgconnor (he/him)Damn enterprise-grade features ;(01:06:36
@connorbaker:matrix.orgconnor (he/him) * Damn the auto-add to project is an enterprise-grade feature ;(01:06:53
@ss:someonex.netSomeoneSerge (back on matrix)NixOS and GH do have some communication channel, I guess we could ask01:07:24
@ss:someonex.netSomeoneSerge (back on matrix)https://github.com/NixOS/nixos-foundation/issues/25#issuecomment-145911322401:13:27
@ss:someonex.netSomeoneSerge (back on matrix)We do have the automation we don't need xD https://github.com/NixOS/nixpkgs/issues/217913#event-fa1e4cc9-b97c-53b4-bf98-595b6708f0fa01:18:10
@connorbaker:matrix.orgconnor (he/him)Screenshot 2023-03-07 at 8.20.12 PM.png
Download Screenshot 2023-03-07 at 8.20.12 PM.png
01:20:20
@connorbaker:matrix.orgconnor (he/him)We only have a few of them; the auto-add to project one would be nice to make sure everything which is tagged with CUDA makes its way into the project. These are the ones I see / have enabled so far:01:20:22
@connorbaker:matrix.orgconnor (he/him)Okay I turned on a few more (reopened items are sent to the backlog to be reprioritized; code changes do the same; code review approvals move it to Ready so it can be merged)01:23:00
@ss:someonex.netSomeoneSerge (back on matrix)I think we should prevent issues being closed when their tickets are just moved around on the board01:30:27
@ss:someonex.netSomeoneSerge (back on matrix)
In reply to @ss:someonex.net
We do have the automation we don't need xD https://github.com/NixOS/nixpkgs/issues/217913#event-fa1e4cc9-b97c-53b4-bf98-595b6708f0fa
We haven't merged https://github.com/NixOS/nixpkgs/pull/218035 yet
01:31:01
@connorbaker:matrix.orgconnor (he/him)Ah, would you rather I have deleted that issue from the project? It was closed as a duplicate, so to me it's "done" in the sense that we no longer need to worry about it01:34:39
@ss:someonex.netSomeoneSerge (back on matrix)

was closed as a duplicate,
oh, ok

01:35:29
@ss:someonex.netSomeoneSerge (back on matrix) *

was closed as a duplicate,

oh, ok

01:35:33
@connorbaker:matrix.orgconnor (he/him)Also just learned that if you accidentally reset your commit and force push it turning your PR into an empty diff it's closed automatically01:35:55
9 Mar 2023
@hexa:lossy.networkhexahttps://opensource.googleblog.com/2023/03/openxla-is-ready-to-accelerate-and-simplify-ml-development.html?m=108:36:41
@ss:someonex.netSomeoneSerge (back on matrix)https://github.com/openxla/xla/issues/113:21:07
@connorbaker:matrix.orgconnor (he/him)

Is there a centralized location for docs for CUDA-maintainer related stuff?
Containing answers to (for example):

  • Do we have any infrastructure (like CI) besides cachix?
  • What populates our cachix?
  • What's the storage limit for our cachix (meaning, is the number of derivations we host a result of limited compute, storage, or both)?
  • If it's not CI populating the cache, what's the process for getting permissions to push to it?
15:47:43
@ss:someonex.netSomeoneSerge (back on matrix)I'd CUDA page on NixOS Wiki should be one15:59:20
@ss:someonex.netSomeoneSerge (back on matrix) * I'd say CUDA page on NixOS Wiki should be one15:59:23
@ss:someonex.netSomeoneSerge (back on matrix)https://nixos.wiki/wiki/CUDA15:59:31
@ss:someonex.netSomeoneSerge (back on matrix)Hasn't been maintained for a while15:59:44
@connorbaker:matrix.orgconnor (he/him)Is there any sort of VCS or approval/review process for that?19:50:45
10 Mar 2023
@connorbaker:matrix.orgconnor (he/him)

These results are preliminary but it looks like using -Xfatbin=-compress-always nearly cut the size of the magma NAR in half (from 429.4M to 233.6M) when building for just 8.6: https://github.com/NixOS/nixpkgs/pull/220402

That's super impressive, so I'm excited to see what it looks like when targeting multiple capabilities.

00:47:45
@ss:someonex.netSomeoneSerge (back on matrix)Just a wiki02:39:29
@connorbaker:matrix.orgconnor (he/him)

Looks like PyTorch nightlies are using a much newer version of Triton relative to what they were a month or two ago. They’ve got their own branch which closely tracks master: https://github.com/openai/triton/tree/torch-inductor-stable.

Packaging that could be difficult given they pull in their own build of MLIR (based on LLVM 17: https://github.com/openai/triton/blob/2c32f4399986045ff25cae201ed3b16d922a9d3b/python/setup.py#L72) and unconditionally grab NVCC from conda (https://github.com/openai/triton/blob/2c32f4399986045ff25cae201ed3b16d922a9d3b/python/setup.py#L107).

We don’t have MLIR packaged yet: https://github.com/NixOS/nixpkgs/pull/163878 (although I think we do build it for ROCm?)

Thoughts?

14:29:53

There are no newer messages yet.


Back to Room ListRoom Version: 9