Meta has just released the latest version of its Llama family of artificial intelligence models including the largest version to date, coming in at 820GB.
The company says this is a new upgrade across the family, calling them Llama 3.1, including reasoning improvements to the small and mid-tier versions.
These will also still be open-source, meaning any company or organization can download, fine-tune, and run the model on its own hardware.
Meta says Llama 3.1 405b, the largest of the family, can match models from OpenAI and Google. This includes reasoning and coding capabilities.
It is available to use on the Groq Cloud inference platform, Meta's own MetaAI chatbot and developer clouds like AWS, IBM and Azure.
What is different about Llama 3.1?
Llama 3.1 405b, which means 405 billion parameters, is the big change for both Meta and the open-source AI community with the company claiming it beats Claude 3.5 Sonnet and GPT-4o on a number of key benchmarks.
While Meta will use this in its own MetaAI chatbot, the model will bring advanced frontier-grade intelligence to companies to install on their own hardware, to adapt to their own needs and to use without paying per-token charges to OpenAI.
Starting today, open source is leading the way. Introducing Llama 3.1: Our most capable models yet.Today we’re releasing a collection of new Llama 3.1 models including our long awaited 405B. These models deliver improved reasoning capabilities, a larger 128K token context… pic.twitter.com/1iKpBJuReDJuly 23, 2024
During training and subsequent fine-tuning Meta says it focused on 'helpfulness' of its chatbot. Explaining in a statement: "With Llama 3.1 405B, we strove to improve the helpfulness, quality, and detailed instruction-following capability of the model in response to user instructions while ensuring high levels of safety. "
It comes with a much larger 128,000 token context window, which means brings it roughly in-line with the industry standard.
What are the Llama 3.1 versions?
Llama 3.1 comes in in three sizes. The first is 8B, which is light-weight and ultra-fast, able to run anywhere including on a smartphone.
Then you have 70B, which is the high performant but cost effective model, mainly run in a data center but could operate on a good gaming PC.
Finally, there’s the new flagship foundation model 405B, which is approaching the same quality and capabilities as the big private models from the likes of OpenAI, Google and Anthropic.
How do you access Llama 3.1?
Llama 3.1, including the new 405b version, is available in the MetaAI chatbot but that has a limited release globally. More widely, you can access it using the Groq Cloud or even the very quick Groq chatbot interface.
Other developer-focused platforms from Microsoft, Google, Amazon, DataBricks and more are also loading the new models and if you've got your own data center you can download all three versions from GitHub and run them yourself.
The two smaller versions — 8b and 70B — are available on Ollama, an easy installer for local AI models on Windows, Mac and Linux.
Meta CEO Mark Zuckerberg predicts MetaAI will be the most widely used assistant by the end of this year, beating out ChatGPT as the number one way to access AI.
He says this will be thanks to expanded access to the assistant through WhatsApp, Instagram, the Ray-Ban smart glasses and the standalone MetaAI chatbot. It is also coming to the Quest mixed-reality headsets to replace the current voice commands.