Our Mission: Build the world's best small language models

Tiny Labs is a community-driven research lab on a mission to build the world's first sub-billion parameter language model to surpass 80% accuracy on the MMLU benchmark. This performance is currently only achieved by models with trillions of parameters.

Lightning Fast

Real-time inference on any device

On Device

Run on phone, laptop, edge devices

Eco-Friendly

1000x less energy consumption

Affordable

Fraction of the cost to train and deploy

Why Small Models Matter

By focusing on small, efficient language models, we're creating a more accessible and sustainable AI ecosystem that empowers individuals and organizations to build sophisticated AI applications without the massive computational resources typically required.

Our Progress

Each month we release a new version of our model, accompanied by a detailed blog post explaining what changed and our progress. Our models improve continuously as we iterate on architecture, training data, and optimization techniques. Each benchmark represents our latest model's performance.

MMLU

Target: 80%
Tiny Labs (July 2025): 40%

Arc Challenge

Target: 70%
Tiny Labs (July 2025): 35%

GSM8K

Target: 60%
Tiny Labs (July 2025): 15%

HumanEval

Target: 50%
Tiny Labs (July 2025): 5%

Inspired by EleutherAI and Marin, we're building a community-led approach to developing small language models. Anyone with a great idea can test it, refine it with peers, and get credit for their impact. Our mission is to make contributing to frontier AI as straightforward as contributing to open-source software.

Community-Driven AI Research for Efficient Language Models

Our Workflow: How Tiny Labs Works

Propose an Experiment

Submit your innovative ideas and detailed proposals via our structured GitHub issues. Focus on architectural changes, optimiser tweaks, or data variations. No idea is too small!

Community Prioritisation

On a bi-weekly/monthly cycle, the community and sponsors vote on the most promising proposals through an open auction system.

Rigorous Evaluation

Selected experiments are evaluated against general benchmarks like MMLU, and potentially sponsor-defined tests, ensuring robust assessment.

Recognition & Rewards

Receive financial rewards, build a transparent reputation, and gain co-authorship on our monthly technical reports. Top contributions are merged into our main model. Small targeted changes (e.g., a new pre-processing step, quality filter, or hyper-parameter tweak) are fairly credited alongside larger architectural overhauls.

Latest Publications

See what we've been up to.

Who we are

We are Pietro and Richard, two Cambridge PhD grads. During our PhD, we shared an office; now we share the same passion for building the world's best small language models.

Pietro Lesci

Pietro Lesci

Co-Founder

Richard Diehl Martinez

Richard Diehl Martinez

Co-Founder