Skip to main content

Meta unveils Llama 3.1, its biggest and best open source model yet

llama 3.1 logo
Meta

Facebook parent company Meta announced the release of its Llama 3.1 open source large language model on Tuesday. The new LLM will be available in three sizes — 8B, 70B, and 405B parameters — the latter being the largest open-source AI built to date, which Meta CEO Mark Zuckerberg describes as “the first frontier-level open source AI model.”

“Last year, Llama 2 was only comparable to an older generation of models behind the frontier,” Zuckerberg wrote in a blog post Tuesday. “This year, Llama 3 is competitive with the most advanced models and leading in some areas. Starting next year, we expect future Llama models to become the most advanced in the industry.”

llama 3.1-405B benchmarks
Meta

Trained on 15 trillion tokens using 16,000 H100 GPUs, Meta claims that the 405B model is significantly larger than its Llama 3 predecessor. It reportedly rivals today’s top closed source models, such as OpenAI’s GPT-4o, Google’s Gemini 1.5, or Anthropic’s Claude 3.5 in “general knowledge, math, tool use, and multilingual translation. Zuckerberg predicted on Instagram on Tuesday that Meta AI would surpass ChatGPT as the most widely used AI assistant by the end of the year.

Recommended Videos

The company notes that all three versions of Llama 3.1 will enjoy expanded prompt lengths of 128k tokens, enabling users to provide added context and up to a book’s worth of supporting documentation. They’ll also support eight languages at launch. What’s more, Meta has amended its license agreement to allow developers to use Llama 3.1 outputs to train other models.

Meta also announced that it is partnering with more than a dozen other companies in the industry to further develop the Llama ecosystem. Amazon, Databricks, and Nvidia will launch full-service software suites to help developers fine-tune their own models based off Llama, while the startup Groq has “built low-latency, low-cost inference serving” for the new family of 3.1 models, Zuckerberg wrote.

Being open-source, Llama 3.1 will be available on all the major cloud services including AWS, Google Cloud, and Azure.

Andrew Tarantola
Andrew Tarantola is a journalist with more than a decade reporting on emerging technologies ranging from robotics and machine…
Meta’s new AI model can turn text into 3D images in under a minute
an array of 3D generated images made by Meta 3D Gen

Meta's latest foray into AI image generation is a quick one. The company introduced its new "3D Gen" model on Tuesday, a "state-of-the-art, fast pipeline" for transforming input text into high-fidelity 3D images that can output them in under a minute.

What's more, the system is reportedly able to apply new textures and skins to both generated and artist-produced images using text prompts.

Read more
Why Llama 3 is changing everything in the world of AI
Meta AI on mobile and desktop web interface.

In the world of AI, you've no doubt heard about what OpenAI and Google have been up to. And now, Meta's Llama LLM (large language model) is becoming an increasingly important player in the game, especially with its open-source nature. Meta recently made a big splash with the launch of its Llama 3 AI model, and it's shaken up the field dramatically.

The reasons why are multiple and varied. It's free to use, it has a wide user base, and yes, it's open source, to name but a few. Here's why Llama 3 is taking the AI industry by storm and may shape its future for some time to come.
Llama 3 is really good
We can debate until the cows come home about how useful AIs like ChatGPT and Llama 3 are in the real world -- they're not bad at teaching you board game rules -- but the few benchmarks we have for how capable these AI are give Llama 3 a distinct advantage.

Read more
OpenAI’s GPT-3 algorithm is here, and it’s freakishly good at sounding human
GPT-2 AI Text Generator

When the text-generating algorithm GPT-2 was created in 2019, it was labeled as one of the most “dangerous” A.I. algorithms in history. In fact, some argued that it was so dangerous that it should never be released to the public (spoiler: It was) lest it ushers in the “robot apocalypse." That, of course, never happened. GPT-2 was eventually released to the public, and after it didn't destroy the world, its creators moved on to the next thing. But how do you follow up the most dangerous algorithm ever created?

The answer, at least on paper, is simple: Just like the sequel to any successful movie, you make something that’s bigger, badder, and more expensive. Only one xenomorph in the first Alien? Include a whole nest of them in the sequel, Aliens. Just a single nigh-indestructible machine sent back from the future in Terminator? Give audiences two of them to grapple with in Terminator 2: Judgment Day.

Read more