
‘Tis the week for small AI models, it seems.
Nonprofit AI research institute Ai2 on Thursday released Olmo 2 1B, a 1-billion-parameter model that Ai2 claims beats similarly-sized models from Google, Meta and Alibaba on several benchmarks. Parameters, sometimes referred to as weights, are the internal components of a model that guide its behavior.
Olmo 2 1B is available under a permissive Apache 2.0 license on AI dev platform Hugging Face. Unlike most models, Olmo 2 1B can be replicated from scratch, as Ai2 has provided the code and data sets (Olmo-mix-1124 and Dolmino-mix-1124) used to develop it.
Small models might not be as capable as their behemoth counterparts, but importantly, they don’t require beefy hardware to run. That makes them much more accessible for developers and hobbyists contending with the limitations of lower-end hardware and consumer machines.
There’s been a raft of small model launches over the past few days, from Microsoft’s Phi 4 reasoning family to Qwen’s 2.5 Omni 3B. Most of these, including Olmo 2 1B, can easily run on a modern laptop or even a mobile device.
Ai2 says Olmo 2 1B was trained on a data set of 4 trillion tokens from publicly available, AI-generated, and manually created sources. Tokens are the raw bits of data that models ingest and generate, with a million tokens equivalent to about 750,000 words.
On a benchmark measuring arithmetic reasoning, GSM8K, Olmo 2 1B scores better than Google’s Gemma 3 1B, Meta’s Llama 3.2 1B, and Alibaba’s Qwen 2.5 1.5B. Olmo 2 1B also eclipses the performance of those three models on TruthfulQA, a test for evaluating factual accuracy.
Techcrunch event
Berkeley, CA
|
June 5
Ai2 has warned that Olmo 2 1B carries risks, however. Like all AI models, it can produce “problematic outputs,” including harmful and “sensitive” content, the organization said, as well as factually inaccurate statements. For these reasons, Ai2 recommends against deploying Olmo 2 1B in commercial settings.