Cohere for AI Releases Open-Source Aya Vision Models for Computer Vision-Based Tasks

Date:

- Advertisement -


Cohere For AI, the firm’s open research division, released new state-of-the-art (SOTA) vision models on Tuesday. Dubbed Aya Vision, the artificial intelligence (AI) models are available in two parameter sizes. The company’s latest frontier models address the inconsistent performance of existing large language models (LLMs) across different languages, especially for multimodal tasks. Aya Vision models can generate outputs in 23 languages and can perform both text-based and image-based tasks. However, it cannot generate images. Cohere has made the AI models available on open-source repositories as well as via WhatsApp.

Cohere Releases Aya Vision AI Models

In a blog post, the AI firm detailed the new vision models. Aya Vision is available in 8B and 32B parameter sizes. These models can generate text, translate text and images across 23 languages, analyse images and answer queries about them, as well as caption images. Both models can be accessed via Cohere’s Hugging Face page and on Kaggle.

Additionally, general users can try out Cohere’s models via a dedicated WhatsApp chat account that can be accessed here. The company says the Aya Vision models are useful for instances when people come across images or artworks they would like to learn more about.

Based on the company’s internal testing, the Aya Vision 8B model outperforms Qwen2.5-VL 7B, Gemini Flash 1.5 8B, and Llama 3.2 11B Vision models on the AyaVisionBench and m-WildVision benchmarks. Notably, the AyaVisionBench benchmark was also developed by Cohere, and its details have been shared in the public domain.

Coming to the Aya Vision 32B model, the company claimed that it outperformed Llama 3.2 90B Vision and Qwen2-VL 72B models on the same benchmarks.

To achieve frontier performance, Cohere claimed that several algorithmic innovations were developed. The Aya Vision models were fed synthetic annotations, developers scaled up multilingual data through translation and rephrasing, and multiple multimodal models were merged in separate steps. The developers observed that in each step, the performance was significantly improved.

Notably, developers can access the open weights of the Aya Vision models from Kaggle and Hugging Face, however, these models are available with a Creative Commons Attribution Non Commercial 4.0 license. It allows for academic and research-based usage but prohibits commercial use cases.



Source link

- Advertisement -

Top Selling Gadgets

LEAVE A REPLY

Please enter your comment!
Please enter your name here

nine + nine =

Share post:

Subscribe

Popular

More like this
Related

Just a moment…

https://www.yorkshireeveningpost.co.uk/sport/football/leeds-united/leeds-united-predicted-lineup-vs-portsmouth-revealed-as-daniel-farke-reacts-to-west-brom-disappointment-5021781Source link

Big Tech Opposes YouTube Exemption from Australia’s Ban on Social Media for Children

Tech giants Facebook and Instagram owner Meta Platforms,...

New Zealand lose key player for Champions Trophy Final against India

His place in New Zealand's side has been...

Here’s Why Apple is Unlikely to Release an M4 Ultra Chip for Macs

The new Mac Studio is available with mismatched...

Top Selling Gadgets