Marin

Logo

Developing models together, openly

GitHub
Discord
Documentation
DeepWiki
HuggingFace
WandB
Speedrun Leaderboard

Marin

Marin is an open lab for building foundation models—together. We’re training powerful models from scratch, and sharing and programmatically documenting every step: the code, the data, the experiments, the mistakes…all in real-time. We invite anyone who shares our vision of open science and open-source to join and contribute, whether you want to try out a new architecture, training algorithm, dataset, evaluation…there is a lot to do!

News (2025-05-19): Read our announcement!

Want to jump in? Install the Marin code and run your first experiment!

Experiments 🧪

Building a foundation model requires countless experiments trying out endless variants of algorithms and datasets. All the experiments we’re doing are captured as GitHub issues (here is a summary).

Here’s the lifecycle of an experiment:

  1. A GitHub issue is created to preregister the experiment (hypotheses, goal).
  2. A pull request is created with code that reproduces the experiment.
  3. Code defines a provenance graph which is executed; results are summarized in a WandB report.

Some examples:

  1. Experiment 935: How does z-loss impact loss?
    [issue, PR, code, execution, WandB]
  2. Experiment 950: How does pretraining learning rate impact SFT?
    [issue, PR, code, execution, WandB]
  3. Experiment 163: Is BERT a better quality filter than fastText?
    [issue, PR, code, execution, WandB]
  4. Experiment 1290: Which optimizers actually outperform AdamW?
    [issue, PR, code, execution, WandB]
  5. Experiment 1183: Are MoEs really better than dense models?
    [issue, PR, code, execution, WandB]
  6. Experiment 702: How should you train on rare task-relevant data?
    [issue, PR, code, execution, WandB]

Models 🌐

We trained some models in Marin:

  1. Marin-8B-Base (deeper-starling) [issue, code, execution, WandB]: Beats Llama 3.1 8B base on 14/19 standard benchmarks! Read more in our retrospective.
  2. Marin-8B-Instruct (deeper-starling-05-15) [execution]: Try it out on Together AI!
  3. Marin-32B-Base [issue, execution]: It is still training! You can watch it live on WandB🍿!

Speedrun 🏃

Have a new architecture or training procedure that you think is more efficient? Participate in the Marin speedrun competition (inspired by the nanogpt speedrun), pick your compute budget, and create the fastest method to train a model to a certain quality! Here’s an example submission. We will offer free compute to scale up top performers. Get started here.

Datashop 🛠️

Want to add new capabilities to the Marin models? Visit our datashop, where you can upload a dataset or craft a prompt to curate a relevant dataset for your task.

For example, we used Llama 3 70B to filter for mathematical educational data (like FineMath). [issue, PR, code, execution]

Acknowledgements

Marin wouldn’t be possible without the generous support of the Google TPU Research Cloud program. We also benefit immensely from the broader open ecosystem, who have released numerous tools and datasets.