Generative AI Publication

Generative AI Publication

Share this post

Generative AI Publication
Generative AI Publication
Grok-1 Is Here: It's The Largest Open-Source LLM With 300B Parameters
Copy link
Facebook
Email
Notes
More

Grok-1 Is Here: It's The Largest Open-Source LLM With 300B Parameters

Elon Musk’s artificial intelligence company, xAI, has released the weights and architecture of their 314 billion parameter Mixture-of-Experts model, Grok-1, under the Apache 2.0 license.

Jim Clyde Monge's avatar
Jim Clyde Monge
Mar 18, 2024
∙ Paid

Share this post

Generative AI Publication
Generative AI Publication
Grok-1 Is Here: It's The Largest Open-Source LLM With 300B Parameters
Copy link
Facebook
Email
Notes
More
1
Share

Elon Musk’s artificial intelligence company, xAI, has released the weights and architecture of their 314 billion parameter Mixture-of-Experts model, Grok-1, under the Apache 2.0 license.

This follows a pledge by Musk last Monday to make Grok freely available to the public. As someone who has been closely following developments in the AI space, I have to say this is a massive step forward in terms of openness and accessibility.

What is Grok?

Grok is a huge language model with 314 billion parameters, making it the largest open-source model currently available. For context, that’s more than double the size of OpenAI’s GPT-3, which was considered a breakthrough when it was released in 2020.

In a chart shared by X user Andrew Kean Gao, you can see how huge Grok’s size is compared to its competitors.

Keep reading with a 7-day free trial

Subscribe to Generative AI Publication to keep reading this post and get 7 days of free access to the full post archives.

Already a paid subscriber? Sign in
© 2025 Jim Clyde Monge
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share

Copy link
Facebook
Email
Notes
More