September 25th, 2024

Llama 3.2 released: Multimodal, 1B to 90B sizes

Llama 3.2 has been released as an open-source AI model in various sizes for text and image processing, enhancing application development and gaining significant traction with over 350 million downloads.

Read original articleLink Icon
Llama 3.2 released: Multimodal, 1B to 90B sizes

Llama 3.2 has been released as an open-source AI model, offering various sizes including 1B, 3B, 11B, and 90B. The 1B and 3B models are designed for multilingual text processing, while the 11B and 90B models can handle both text and image inputs. This release aims to enhance the development of efficient applications, particularly for on-device use, such as summarizing discussions or processing images. The Llama Stack provides a streamlined developer experience, allowing for quick application development in multiple programming languages, including Python and Swift. The models have been evaluated against over 150 benchmark datasets, demonstrating strong performance in various tasks. Llama models have gained significant traction, with over 350 million downloads on Hugging Face, indicating their popularity in the open-source community. Partnerships with companies like ARM, MediaTek, and Dell are facilitating the integration of Llama models into mobile and edge devices, promoting the adoption of AI in enterprise applications. The Llama Stack is positioned to simplify AI application development across diverse use cases, enhancing productivity and collaboration in various industries.

- Llama 3.2 offers models in multiple sizes for text and image processing.

- The Llama Stack simplifies application development and supports various programming languages.

- Over 350 million downloads highlight the popularity of Llama models in the open-source community.

- Partnerships with tech companies enhance the deployment of Llama models on mobile and edge devices.

- The release aims to improve AI application development across various enterprise use cases.

Link Icon 8 comments
By @wesleyyue - about 2 months
Interesting observations:

* Llama 3.2 multimodal actually still ranks below Molmo from ai2 released this morning.

* AI2D: 92.3 (3.2 90B) vs 96.3 (of Molmo 72B)

* Llama 3.2 1B and 3B is pruned from 3.1 8B so no leapfrogging unlike 3 -> 3.1.

* Notably no code benchmarks. Deliberate exclusion of code data in distillation to maximize mobile on-device use cases?

Was hoping there would be some interesting models I can add to https://double.bot but doesn't seem like any improvements to frontier performance on coding.

By @ChrisArchitect - about 2 months
By @jarbus - about 2 months
I’m more excited about Llama stack, I can’t wait for local models to be able to use tools in a standard way.
By @fulladder - about 2 months
When will it come to ollama? That's my preferred quantization platform.
By @artninja1988 - about 2 months
Are users in the EU not allowed to use it, like they threatened to do so recently?
By @oriettaxx - about 2 months
do you have an idea how long will take to have it available in ollama ?
By @pheeney - about 2 months
What is the best provider for API use of llama frontier models considering pricing / reputation?