| 10 Dec 2025 |
Qyriad | In reply to @piegames:flausch.social Due to unfortunate design decisions made (or rather, not made) back when I was in Kindergarten, parsing is on the critical path for evaluation time we would like to see benchmarks on this… we'd honestly sooner suspect filesystem IO as a bottleneck over parse-time | 17:18:48 |
Qyriad | In reply to @piegames:flausch.social Due to unfortunate design decisions made (or rather, not made) back when I was in Kindergarten, parsing is on the critical path for evaluation time * we would like to see benchmarks/profiles on this… we'd honestly sooner suspect filesystem IO as a bottleneck over parse-time | 17:19:17 |
Qyriad | (having looked at profiles of eval before, but not in a while) | 17:19:29 |
piegames | no, parsing is super fast, mostly thanks to horrors, it's just that Nixpkgs is 3.6MLOC over 40k files and every single NixOS eval loads a good chunk of that, realistically multiple times. The main issue is that caching was never a thought in the architecture and adding it afterwards is really tricky | 17:21:27 |
Qyriad | don't forget that every single Nixpkgs package loads hundreds if not thousands of other nix files | 17:22:24 |
piegames | yes, we're close enough to that atm, there still is one full AST walk for bindVars that's still costly (I tried to remove it with horrors but unexplicably did not make anything faster) and general AST allocation cost (more bump allocators would help a lot there probably) | 17:22:40 |
Qyriad | In reply to @qyriad:katesiria.org don't forget that every single Nixpkgs package loads hundreds if not thousands of other nix files (and instantiates half a gazillion derivations…) | 17:22:52 |
piegames | we won't get much faster than that (though I expect that a parse that can directly emit Bytecode should be a little bit faster still because more compact representation), but the Rust rewrite still needs to be at least as fast as now and that's no small feat | 17:23:20 |
piegames | * we won't get much faster than that (though I expect that a parser that can directly emit Bytecode should be a little bit faster still because more compact representation), but the Rust rewrite still needs to be at least as fast as now and that's no small feat | 17:23:45 |
Rutile (Commentator2.0) feel free to ping | In reply to @piegames:flausch.social we won't get much faster than that (though I expect that a parser that can directly emit Bytecode should be a little bit faster still because more compact representation), but the Rust rewrite still needs to be at least as fast as now and that's no small feat Could the rust version be possibly be writtten with cache in mind from the beginning? | 17:25:15 |
piegames | yes, but the issue is, what is your cache key? | 17:25:37 |
piegames | inode number and ctime might be our best bet | 17:28:31 |
piegames | but in terms of cachable data structures, bytecode gives us that for free | 17:29:02 |
piegames | * but in terms of cachable data structures, bytecode gives us that bit for free | 17:29:28 |
Charles | In reply to @piegames:flausch.social yes, but the issue is, what is your cache key? What do other interpreters do? | 17:31:54 |
piegames | I don't think any other interpreters have comparable performance requirements for parsing | 17:33:30 |
piegames | if there constraints weren't as tight we could just go content-addressed and use a hash | 17:33:43 |
Qyriad | bytecode is our::Qyriad's long term goal | 17:35:18 |
Qyriad | or really whenever someone pays us for it | 17:35:25 |
Qyriad | * or whenever someone pays us for it | 17:35:36 |
piegames | I think browser engines have sufficient control over their baseline caching pipelines and what data changes when that they can probably just track what changes easily, but I'm guessing here | 17:36:08 |
KFears (burnt out) | In reply to @piegames:flausch.social no, parsing is super fast, mostly thanks to horrors, it's just that Nixpkgs is 3.6MLOC over 40k files and every single NixOS eval loads a good chunk of that, realistically multiple times. The main issue is that caching was never a thought in the architecture and adding it afterwards is really tricky I don't know how many of those lines are going to be actually parsed (now I wish I could somehow learn that into), but assuming 20% utilization (that doesn't sound too unreasonable, considering that my desktop has 2k packages, but there's a lot of re-evals due to module system, splicing, package variations etc.), that doesn't sound too terrible? Like, other interpreters and compilers probably deal with 700k LOC codebases. How do they deal with that? | 17:36:40 |
piegames | I'm dreaming of tackling it for 2.96 really | 17:36:58 |
piegames | I already have a half-working prototype from a year ago, horrors had one too, and we should soon have all the parts together for doing it | 17:37:37 |
piegames | main question is whether to do it in C++ or to block it on RPC and do it in Rust | 17:37:53 |
helle (just a stray cat girl) | please don't do it in C++ | 17:38:13 |
helle (just a stray cat girl) | we don't want to have to migrate that later on | 17:38:26 |
Qyriad | we would either block it on RPC, or bind it to existing libexpr over C bindings. the refactoring necessary for libexpr to use a bytecode interpreter over C ABI would be like half the battle for the decoupling we need to do anyway | 17:40:46 |
Sergei Zimmerman (xokdvium) | FWIW v8 uses the source code hash as the cache key https://github.com/v8/v8/blob/427f7cce6d69a2d6ce113200e8dcc1151765058c/src/snapshot/code-serializer.cc#L820-L834 | 17:41:09 |
Sergei Zimmerman (xokdvium) | So the closest thing would be to hash .nix files -> cached bytecode -> cached optimizer -> .... -> profit? | 17:42:32 |