Tech »  Topic »  Meta’s Llama 3.2 launches with vision to rival OpenAI, Anthropic

Meta’s Llama 3.2 launches with vision to rival OpenAI, Anthropic


Meta’s large language models (LLMs) can now see. 

Today at Meta Connect, the company rolled out Llama 3.2, its first major vision models that understand both images and text. 

Llama 3.2 includes small and medium-sized models (at 11B and 90B parameters), as well as more lightweight text-only models (1B and 3B parameters) that fit onto select mobile and edge devices.

“This is our first open-source multimodal model,” Meta CEO Mark Zuckerberg said in his opening keynote today. “It’s going to enable a lot of applications that will require visual understanding.”

Like its predecessor, Llama 3.2 has a 128,000 token context length, meaning users can input lots of text (on the scale of hundreds of pages of a textbook). Higher parameters also typically indicate that models will be more accurate and can handle more complex tasks. 

Meta is also today for the first time sharing ...


Copyright of this story solely belongs to venturebeat . To see the full text click HERE