NixOS CUDA | 288 Members | |
| CUDA packages maintenance and support in nixpkgs | https://github.com/orgs/NixOS/projects/27/ | https://nixos.org/manual/nixpkgs/unstable/#cuda | 58 Servers |
| Sender | Message | Time |
|---|---|---|
| 20 Feb 2023 | ||
In reply to @connorbaker:matrix.orgThat's always tough. I think your approach here, all in one PR but with clear commits is fine. The thing is, there are just not that many contributors in Nixpkgs to CUDA and it's quite a big PR so it takes some effort to review. | 18:47:27 | |
| https://github.com/NixOS/nixpkgs/pull/217322 is closer to what I envision doing once I've further split apart that large PR. Does that seem okay? Is there an automated tool I should be using to do something similar to this? | 19:03:27 | |
Thank you all for the work you do maintaining the CUDA-accelerated packages. Building jaxlib and tensorflowWithCuda repeatedly is awful. | 21:45:11 | |
| 21 Feb 2023 | ||
| https://github.com/NixOS/nixpkgs/pull/217497 | 16:55:31 | |
| Has anyone used or set up CCACHE for any of the CUDA derivations? I know they take a while to build and I'm curious what's been done to try to reduce build times | 21:07:03 | |
| Hi! Thank you for investing your time and work in this right now! | 21:20:45 | |
Interesting. I just looked up the ccache nixos wiki page, it suggests one can just drop in something called ccacheStdenv | 21:21:46 | |
Do you know if packages built that way would work as substitutes for normal stdenv ones? | 21:22:59 | |
| Unfortunately I think it would be a different derivation :l | 21:35:02 | |
| so, no magic( | 21:57:53 | |
| From a conversation I had, it seems like it's intended more for use as a dev-shell than for the end derivation | 23:00:04 | |
| 22 Feb 2023 | ||
| so, I have an application that wants tflite_runtime | 00:17:18 | |
| can I substitute that with calls to actual tensorflow? | 00:17:31 | |
| not keen on packaging tflite | 00:18:48 | |
| I'm not familiar with tflite, but good news is it seems that they build it with cmake, not bazel | 08:56:05 | |
| https://www.tensorflow.org/lite/guide/build_cmake | 08:57:43 | |
| tensorflow.life is a thing | 09:38:47 | |
| So no building required i think | 09:39:09 | |
| * tensorflow.lite is a thing | 13:24:36 | |
| 13:24:53 | |
| this is mostly what I did | 13:24:57 | |
| Is there a recommended way to get in touch with NVIDIA about their docs? For example, https://docs.nvidia.com/cuda/archive/11.0.3/ gives me an access denied, and some of their tables in their older docs are missing supported compute capabilities (https://docs.nvidia.com/cuda/archive/11.2.1/cuda-compiler-driver-nvcc/index.html#gpu-feature-list vs https://docs.nvidia.com/cuda/archive/11.3.1/cuda-compiler-driver-nvcc/index.html#gpu-feature-list, sm_37 reappears, but sm_52 is missing in both) | 15:05:30 | |
| Ah, the link for their 11.0.x docs on https://developer.nvidia.com/cuda-toolkit-archive is wrong -- it follows the 10.2 format so it should be something like https://docs.nvidia.com/cuda/archive/11.0/cuda-compiler-driver-nvcc/index.html#gpu-feature-list | 15:09:01 | |
| 23 Feb 2023 | ||
| If anyone has any knowledge to contribute, I'd appreciate it: https://github.com/NixOS/nixpkgs/issues/217780 | 01:14:30 | |
| RE: Getting in touch, I'd recommend starting a new thread in https://forums.developer.nvidia.com/c/8 | 03:09:29 | |
| NVCC has a certain range of compilers it supports. I know that currently we export CC/CXX/CUDAHOSTCXX as appropriate to handle that... but that only changes things in the current derivation. Since the default language standard (like c++11 -> c++14) can change between compiler releases, it's possible that we build a derivation with an NVCC-supported version of GCC or clang, but the libraries that derivation links against were built with a different compiler version with a different language standard. That can manifest as missing or broken symbols during linking, right? | 21:51:31 | |
| 24 Feb 2023 | ||
| Example of me trying to run something I just packaged (https://github.com/connorbaker/bsrt) and maybe getting bitten by (what I think is) exactly this:
OpenCV specifies the CUDA host compiler, but does not set the C or C++ compilers. I'm trying a build with a patched derivation for opencv and hoping that resolves the problem. (Also, OpenCV apparently doesn't build for specific GPU architectures or take advantage of CUDNN!) | 01:02:04 | |
It did! Now I'm seeing a different error of RuntimeError: CUDA driver error: PTX JIT compiler library not found, but that's progress :) | 01:32:45 | |
* It did! Now I'm seeing a different error of RuntimeError: CUDA driver error: PTX JIT compiler library not found, but that's because I'm not using nixGL yet on a non-NixOS machine | 01:41:42 | |
| Is there an issue with
but on master I see
(might be missing something because I'm not immediately finding workarounds that were necessary for other CUDA packages since gcc was bumped) | 19:06:53 | |