LinkedIn and 3rd parties use essential and non-essential cookies to provide, secure, analyze and improve our Services, and to show you relevant ads (including professional and job ads) on and off LinkedIn. Learn more in our Cookie Policy.
Select Accept to consent or Reject to decline non-essential cookies for this use. You can update your choices at any time in your settings.
We are a Seattle-based non-profit AI research institute founded in 2014 by the late Paul Allen. We develop foundational AI research and innovation to deliver real-world impact through large-scale open models, data, robotics, conservation, and beyond.
Artificial Intelligence, Deep Learning, Natural Language Processing, Computer Vision, Machine Reading, Machine Learning, Knowledge Extraction, Common Sense AI, Machine Reasoning, Information Extraction, and Language Modeling
We just added support for the new Olmo3 models directly on HuggingFace—making it a little bit easier for everyone to test and deploy truly open-source AI.
The Public AI Inference Utility now supports:
Olmo-3-32B-Think (via Parasail): https://lnkd.in/e8Jcdzb3
Olmo-3-7B-Instruct (via our partners at Intel/AWS 🙏 ): https://lnkd.in/eWjkSu-m
Olmo-3-7B-Think (via Cirrascale): https://lnkd.in/efWA_XS3
This work is part of our broader effort at the Inference Utility (https://publicai.co) to make high-quality, openly licensed models more accessible across the ecosystem.
Congrats to the AI2 team on pushing the field forward with another strong release. Also tagging some of the underlying inference providers (Parasail, Cirrascale) who are helping make the Olmo3 release happen. We see you. 👏👏
Ai2Hugging FaceParasailCirrascale Cloud ServicesIntel CorporationAmazon Web Services (AWS)Kyle Wiggers for AI2 connections
Diego Bailón Humpert for Intel compute hookup
Joseph Low for leading the implementation at Public AI
Joshua Tan for orchestration
Julien ChaumondSimon B. for helping us fix some rate limiting issues on HF
⚠️ Update on Deep Research Tulu (DR Tulu), our post-training recipe for deep research agents: we’re releasing an upgraded version of our example agent, DR Tulu-8B (RL), that matches or beats systems like Gemini 3 Pro & Tongyi DeepResearch-30B-A3B on core benchmarks.
At just 8B params – lightweight enough to run on a single GPU – DR Tulu-8B (RL) delivers high-quality multi-step reasoning & synthesis for complex questions while staying open, highly inspectable, and easy to customize. 🔍
DR Tulu-8B (RL) is also dramatically cheaper per query than other deep research agents. On ScholarQA-CS2, it costs just ~$0.0019/query vs. ~$0.13 for Gemini 3 Pro + Search, ~$0.29 for GPT-5 + Search, ~$1.80 for OpenAI Deep Research, and ~$0.032 for Tongyi DeepResearch-30B-A3B.
→ More info here: https://lnkd.in/eJtgyChR
To make DR Tulu-8B (RL) practical, we’re releasing an inference engine (via CLI) so you can host the model locally and plug in custom search/browsing tools via MCP. We’re also sharing an updated paper on arXiv.
Get started:
💻 Run DR Tulu locally: https://lnkd.in/eK2Csq-2
⬇️ Model: https://lnkd.in/ehQqCuYw
📄 Technical report on arXiv: https://lnkd.in/ezhZgx8j
Our Olmo 3 models are now available via API on OpenRouter, Inc! Try Olmo 3-Instruct (7B) for chat & tool use, and our reasoning models Olmo-3 Think (7B & 32B) for more complex problems. 👉 https://lnkd.in/efRrscke
Today we’re announcing Olmo 3—our leading fully open language model suite built for reasoning, chat, and tool use, & an open model flow that exposes not just the final weights, but the entire training journey.
Most models ship as a single opaque snapshot. Olmo 3 opens the model flow end to end – pretraining, mid-training, and post-training – plus data recipes and code, so you can see how capabilities are built and customize any stage of the process.
Meet the Olmo 3 family:
🏗️ Olmo 3-Base (7B, 32B)—foundations for post-training with strong code, math, and reading comprehension skills
🛠️ Olmo 3-Instruct (7B)–focused on multi-turn chat and tool use
🧠 Olmo 3-Think (7B, 32B)–“thinking” models that surface their reasoning steps
All are compact, dense models designed to run on hardware ranging from laptops to research clusters.
Under the hood, we trained Olmo 3 on ~6T tokens from our new Dolma 3 pretraining dataset, plus new post-training sets with stronger data decontamination and richer math/code/reasoning mixes. A long-context extension pushes Olmo 3’s context window to ~65K tokens—enough for full papers, books, and other long files.
At the center is Olmo 3-Think (32B), the best fully open 32B-scale reasoning model we’re aware of, alongside our strongest 32B base model.
In our evaluations:
⦿ Olmo 3-Think (32B) is the strongest fully open 32B-scale reasoning model
⦿ Olmo 3-Base models beat fully open Marin & Apertus and rival Qwen 2.5 and Gemma 3
⦿ Olmo 3-Instruct (7B) beats Qwen 2.5, Gemma 3, and Llama 3.1 on tough chat + tool-use benchmarks
We’re also rolling out a major Ai2 Playground upgrade alongside Olmo 3:
🤔 Thinking mode to see intermediate reasoning on complex tasks
🧰 Tool calling so you can define JSON-schema tools or call tools via our Asta platform
Olmo 3 is wired into OlmoTrace in the Ai2 Playground, so you don’t just see its behavior—you can trace it. For example, you can ask Olmo 3-Think (32B) to answer a general-knowledge question, then use OlmoTrace to inspect where and how the model may have learned to generate parts of its response.
If you care about AI you can customize, inspect, and improve, Olmo 3 is for you—available now under Apache 2.0.
Watch an interview with Olmo leads Hanna Hajishirzi and Noah Smith about how & why we built Olmo 3 and what comes next 👉 https://lnkd.in/eGHnu6TH
Then, dive deeper & get started:
✨ Try Olmo 3 in the Ai2 Playground → https://lnkd.in/eniFwyWC
💻 Download the models: https://lnkd.in/eMQWZr2q
📝 Read more in our blog: https://lnkd.in/e3vDT25z
📚 Check out the tech report: https://lnkd.in/ek-ucc2Q
Today we’re releasing Deep Research Tulu (DR Tulu)—the first fully open, end-to-end recipe for long-form deep research, plus an 8B agent you can use right away. 🚀
Our DR Tulu recipe enables you to train agents that can plan multi-step research workflows, search across web pages, academic papers, & specialized tools, then synthesize findings into clear explanations with inline citations. Under the hood, DR Tulu agents dynamically switch between web search, browsing, and scholarly tools depending on the research question. 📈
DR Tulu introduces Reinforcement Learning with Evolving Rubrics (RLER), a reward scheme grounded in actual search results that evolves during training to capture new strategies + reduce reward hacking. Our MCP-based inference system lets you bring your own tools to expand DR Tulu’s capabilities.
The goal: make expert-level research more accessible, transparent, and explainable. 🧭📚
Strong performance: Our open DR Tulu-8B (RL) example agent beats other open models and matches or outperforms closed systems like OpenAI Deep Research and Perplexity Deep Research on challenging benchmarks. It adapts to the task, delivering one-line answers for simple questions or detailed reports for complex topics.
Cost-effective: DR Tulu-8B (RL) costs ≤ $0.0075 on our ScholarQA-CSv2 benchmark, compared to ~$1.80 for OpenAI Deep Research & ~$1.30 for our Asta pipeline with a Claude Sonnet backend.
Dive in & learn more:
📚 Blog: https://lnkd.in/eJtgyChR
✏️ Paper: https://lnkd.in/eZJ2pK6W
💻 Models: https://lnkd.in/ehQqCuYw
⌨️ Code: https://lnkd.in/eXfuFNCb
Introducing the OlmoEarth Platform 🌍, state-of-the-art AI paired with ready-to-use open infrastructure to turn Earth data into clear, up-to-date insights.
Now rolling out, OlmoEarth Platform is an open, scalable, end-to-end system that transforms satellite imagery, radar, elevation data, and more into actionable intelligence—maps when helpful, plus change alerts & custom dashboards.
We're releasing:
💻 Code: https://lnkd.in/e8vt45rG
➡️ OlmoEarth models (more info below): https://lnkd.in/eAdQy9pN
📝 A technical report: https://lnkd.in/eVPeNqY5
🌍 The OlmoEarth Platform: https://lnkd.in/en-A8sgS
Updates arrive within hours, not years, and the integrated workflow cuts cost and manual effort, so regular refreshes fit real programs and budgets. Under the hood, our industry-leading OlmoEarth foundation model family fuses multi-sensor Earth data and adapts quickly to local needs—one open model, many missions, fast to fine-tune & deploy.
Learn more about our OlmoEarth models, which top key industry benchmarks and partner use cases for Earth observation, here → https://lnkd.in/eCmmiDFM
By applying AI to a planet’s worth of data, we’re providing governments, NGOs, and communities with timely and trustworthy insights so people can act faster + with confidence to protect both nature and livelihoods. 👇
🌲 Wildfire deployments with NASA Jet Propulsion Laboratory are mapping live fuel moisture at scale to inform readiness.
🌱 International Food Policy Research Institute (IFPRI) in Nandi County, Kenya & Mozambique produced current countywide crop-type maps that provide the insights needed to improve seasonal planning & address food security challenges.
🌊 Global Mangrove Watch is refreshing mangrove baselines faster, with higher accuracy and less manual review by experts, enabling conservationists + governments to respond more quickly to threats to mangroves.
🔎 Amazon Conservation is identifying likely drivers of deforestation using high-resolution satellite scenes and applying a fine-tuned model to classify loss drivers for alerts across Peru, Bolivia, Colombia, and Brazil.
Our mission is to build AI that serves science and society. If you’re working in food security, wildfire resilience, or on sustainability and conservation initiatives – or build tools for those who do – please get in touch. 🤝
Learn more → https://lnkd.in/etW6w6ZG
"Innovation in the Open" is under way at Ai2 HQ! Sharing with our #SeattleAIWeek guests our latest research and hands-on experiences with our cutting-edge tools.
We’re rolling out olmOCR 2—the next major update to our open OCR model for complex documents & scans. 📝
olmOCR 2 turns messy files with tables, equations, handwriting, and more into clean text. Under the hood, we combine synthetic data with unit tests as verifiable rewards to push state-of-the-art performance on challenging docs.
What’s new:
◆ Stronger text recognition: Trained with a new data mix, including 20,000 historical pages for better coverage of aged and degraded materials. Example: olmOCR 2 can now read Abraham Lincoln’s handwriting correctly, recovering the date “January 10th” in his 1864 letter to Major General Hitchcock. ✍️
◆ Big benchmark gains: 82.4 on olmOCR-Bench (up from 78.5), with improvements across every document category. 📈
◆ Faster & cheaper: New FP8 quantized model (olmOCR-2-7B-1025-FP8) reaches ~3,400 output tokens/sec on a single H100—enough to process 10,000 pages for < $2. 🚀
◆ Adapt to your data: Want to fine-tune for your domain? We provide everything you need to customize and deploy. 🔧
Available now, and on the Deep Infra Inc. & Parasail APIs. We’re also updating our demo—try olmOCR 2 today!
📚 Learn more: https://lnkd.in/ecRNFhrR
💻 Model: https://lnkd.in/exFh8zjf
💬 Discuss: https://discord.gg/ai2