What is Speedrun?

Get started with Speedrun

Speedrun is a community-driven initiative by the Marin project to track and optimize the training efficiency of large language models. Have a new architecture or training procedure that you think is more efficient? Participate in the Marin speedrun competition (inspired by the nanogpt speedrun), pick your compute budget, and create the fastest method to train a model to a certain quality!

On this page, you can find leaderboards for different speedrun tracks, each targeting a specific loss threshold. You can click on any run to view the code that generated it, or view the Weights & Biases link for the model! We also track the overall Pareto frontier of models, allowing us to track efficiency-performance tradeoffs across all tracks.

We invite you to join us in the search for more performant and efficient training methods!

Speedrun Tracks

Select a track to view leaderboard results for a specific model run. Each track now corresponds to a particular Llama model size and its best run's BPB.

Select a track to see its description and leaderboard.

Total Runs in Track

-

Best FLOPs in Track

-

Best C4-EN BPB

-

Overall Pareto Frontier: FLOPs vs. C4-EN BPB

Track Leaderboard

Select a track to view its leaderboard

Rank Run Name Author Date Added Model Size* Training Time Total FLOPs* C4-EN BPB W&B Run
* Model size here refers to the total number of trainable parameters
* Total FLOPs here refers to hardware FLOPs performed during training