Artificial Intelligence Learning 🤖🧠🦾

Artificial Intelligence Learning 🤖🧠🦾

Share this post

Artificial Intelligence Learning 🤖🧠🦾
Artificial Intelligence Learning 🤖🧠🦾
Allen Institute for AI - What is OLMo?
Copy link
Facebook
Email
Notes
More

Allen Institute for AI - What is OLMo?

AI2: Allen Institute for AI releases fully open-source large language model

Michael Spencer's avatar
Michael Spencer
Feb 02, 2024
∙ Paid
5

Share this post

Artificial Intelligence Learning 🤖🧠🦾
Artificial Intelligence Learning 🤖🧠🦾
Allen Institute for AI - What is OLMo?
Copy link
Facebook
Email
Notes
More
1
1
Share
AI2 – Allen Institute for AI Vector Logo | Free Download - (.SVG + .PNG)  format - SeekVectorLogo.Com

Hey Everyone,

Just like we predicted it’s open-source AI developments that are taking center stage so far in 2024.

There’s rarely been such an exciting time for Open-source LLMs as is the climate of 2024. From the saga this week of Mistral Miqu to the potential of Llama-3 later this year, it’s hard to keep up with everything.

But let’s give AI2 some credit here.

Nathan Lambert formerly at Hugging Face recently shared:

Today, we’re releasing our first pretrained Open Language Models (OLMo) at the Allen Institute for AI (AI2), a set of 7 billion parameter models and one 1 billion parameter variant. This line of work was probably the main reason I joined AI2 and is the biggest lever I see possible to enact meaningful change in how AI is used, studied, and discussed in the short term.

OLMo will represent a new type of LLM enabling new approaches to ML research and deployment, because on a key axis of openness, OLMo represents something entirely different. OLMo is built for scientists to be able to develop research directions at every point in the development process and execute on them, which was previously not available due to incomplete information and tools.

Depending on the evaluation methods, OLMo 1 is either the best 7 billion parameter base model available for download or one of the best. This relies on a new way of thinking where models are judged on parameter plus token budget, similar to how scaling laws are measured for LLMs.

You can find the core model here: https://lnkd.in/gxkCzPnU
I wrote about it here (personal take): https://lnkd.in/ga5FQgu7
The technical paper is here: https://lnkd.in/gTMgiV-T

OLMo is entirely open source and a cross-institutional collaboration

Share

Keep reading with a 7-day free trial

Subscribe to Artificial Intelligence Learning 🤖🧠🦾 to keep reading this post and get 7 days of free access to the full post archives.

Already a paid subscriber? Sign in
© 2025 Michael Spencer
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share

Copy link
Facebook
Email
Notes
More