!eWOErHSaiddIbsUNsJ:nixos.org

NixOS CUDA

293 Members
CUDA packages maintenance and support in nixpkgs | https://github.com/orgs/NixOS/projects/27/ | https://nixos.org/manual/nixpkgs/unstable/#cuda58 Servers

You have reached the beginning of time (for this room).


SenderMessageTime
31 Dec 2024
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8) also packaging flash attention now because hopefully it supports fp8 training where PyTorch's implementation does not
why does it require so much memory to build? What is NVCC doing?
16:09:28
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8)Screenshot 2024-12-31 at 11.07.56 AM.png
Download Screenshot 2024-12-31 at 11.07.56 AM.png
16:09:33
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8)

Well, that didn't work

[DEBUG    | DotProductAttention]: Disabling FlashAttention as FlashAttention 2 does not support FP8
[DEBUG    | DotProductAttention]: Disabling UnfusedDotProductAttention as it does not support FP8
[DEBUG    | DotProductAttention]: Disabling FusedAttention as no backend supports the provided input
[DEBUG    | DotProductAttention]: Available backends = {FlashAttention=False, FusedAttention=False, UnfusedDotProductAttention=False}
[DEBUG    | DotProductAttention]: Selected backend = NoBackend
16:40:57
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8)Looks like the authors of Flash Attention are looking at support for fp8 (which comes with v3, currently only available for Hopper) for Ada series per their rebuttals on their paper: https://openreview.net/forum?id=tVConYid20&referrer=%5Bthe+profile+of+Tri+Dao%5D%28%2Fprofile%3Fid%3D~Tri_Dao1%2916:53:06
@kaya:catnip.eekaya 𖤐 changed their profile picture.21:48:16
1 Jan 2025
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8) pushed the changes I had locally for nix-cuda-test (https://github.com/ConnorBaker/nix-cuda-test), if anyone wants to play with transformer engine or flash attention. I'll probably work on upstreaming those at some indeterminate point in time, but I don't know if they'll work with what's in-tree right now 03:59:10
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8) * pushed the changes I had locally for nix-cuda-test (https://github.com/ConnorBaker/nix-cuda-test), if anyone wants to play with transformer engine or flash attention (both for PyTorch). I'll probably work on upstreaming those at some indeterminate point in time, but I don't know if they'll work with what's in-tree right now 03:59:24
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8) SomeoneSerge (utc+3): are you aware of a clean, cross-platform way to handle patching the path to libcuda.so (as needed in https://github.com/NixOS/nixpkgs/pull/369495#issuecomment-2566002172)? Is it fair to assume that on non-NixOS systems, whatever wrapper people use (like nixGL or nixglhost) will add libcuda.so to LD_LIBRARY_PATH? 04:00:43
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8)A new hope? https://www.phoronix.com/news/ZLUDA-v4-Released11:14:50
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8)https://github.com/NixOS/nixpkgs/pull/36995613:22:40
@mjolnir:nixos.orgNixOS Moderation Botchanged room power levels.14:26:32
@connorbaker:matrix.orgconnor (burnt/out) (UTC-8)Thanks for the feedback Serge :)20:37:37
5 Jan 2025
@techyporcupine:matrix.org@techyporcupine:matrix.org set a profile picture.20:36:11
7 Jan 2025
@ruroruro:matrix.orgruro joined the room.03:39:47
@ss:someonex.netSomeoneSerge (back on matrix)Hi, sorry if I missed anything, my homeserver was offline for a week21:58:59
Jitsi widget removed by @ss:someonex.netSomeoneSerge (back on matrix)21:59:20

Show newer messages


Back to Room ListRoom Version: 9