Start date: April 9 , 2024
End date: April 9, 2026
Market with a shorter timeline:
inspired from this tweet by Andrej Karpathy:
Btw writing the llm.c training code would imo be a very interesting, impressive, self-contained and very meta challenge for LLM agents.
The prompt is: Take the PyTorch code train_gpt2.py And write, compile and unit test a single .c file that reproduces the training: train_gpt2.c
The current models are not there, but we can check back in a year or two or so. If that worked...
People are also trading
Buying YES at 24%. Karpathy's llm.c already proved the task is well-defined and tractable — the CPU reference implementation is ~1000 lines of clean C. Current frontier models (Claude Opus 4.6, GPT-5.4) are strong at both C programming and understanding ML training pipelines. With agentic coding frameworks now mainstream, someone attempting this challenge in the next 23 days seems likely.
The hardest part isn't the generation — it's the validation. Writing correct backprop and optimizer code in C with matching numerical outputs to PyTorch is error-prone. But with iterative debugging loops, current AI can handle this. My estimate: ~35%.
@Bayesian https://github.com/karpathy/llm.c/blob/master/train_gpt2.py
Note that if we expect a feature equivalent, single file implementation it would have to also include an implementation of FlashAttention, among other things.
@MalachiteEagle
> It straight up one-shotted it
> idk if the implementation is correct
Many such cases