Meta has announced Llama 3.2, its latest open-source AI model designed to facilitate visual and voice interactions, making AI more accessible to users.
Llama 3.2 represents a significant advancement in Meta's AI capabilities, introducing multimodal functionalities that allow the model to process both images and text. This enhancement enables developers to create more versatile applications, including augmented reality experiences, visual search engines, and document analysis tools.
The model includes two vision models with 11 billion and 90 billion parameters, as well as two lightweight text-only models with 1 billion and 3 billion parameters. These smaller models are optimized for mobile devices, allowing for efficient AI performance on platforms with limited resources.
The open-source nature of Llama 3.2 encourages collaboration among developers, researchers, and businesses, fostering innovation and transparency in the AI community. By making these models freely available, Meta aims to accelerate the development of AI applications across various industries.
In addition to Llama 3.2, Meta has introduced new voice capabilities for its AI assistant, Meta AI. The assistant now supports voice input and features celebrity voice options, enhancing user engagement and personalization. These updates are accessible on platforms such as Facebook, Messenger, WhatsApp, and Instagram direct messages.
These developments underscore Meta's commitment to advancing AI technology and making it more accessible and functional for users worldwide.