Home Community Reka AI Releases Reka Flash: An Efficient and Capable State-of-the-Art 21B Multimodal Language Model

Reka AI Releases Reka Flash: An Efficient and Capable State-of-the-Art 21B Multimodal Language Model

0
Reka AI Releases Reka Flash: An Efficient and Capable State-of-the-Art 21B Multimodal Language Model

Reka addresses the necessity for advanced language and vision models with their state-of-the-art multimodal and multilingual language model, Reka Flash. It might probably perform excellently on various benchmarks of LLM even with a smaller model, Reka Edge, with just 7B of trainable parameters. The models solve the challenges of achieving high performance across diverse tasks and languages using limited computational resources.

Existing models like Gemini Pro, GPT-3.5, and Llama-2 70B are discussed as benchmarks for comparison. Reka Flash and Reka Edge offer turbo-class and compact variant models, which leverage pretraining on text from over 32 languages and are evaluated on multiple benchmarks, including language understanding, reasoning, multilingual tasks, and multimodal interactions. Reka Flash competes with leading models on language benchmarks and vision tasks, while Reka Edge targets local deployments with resource constraints.

Reka Flash utilizes instruction tuning and reinforcement learning with Proximal Policy Optimization (PPO) to boost its chat capabilities. Its performance is evaluated in each text-only and multimodal chat domains, and it presents competitive results against models like GPT-4, Claude, and Gemini Pro. Reka Edge, optimized for local deployments, outperforms comparable models equivalent to Llama 2 7B and Mistral 7B on standard language benchmarks, indicating efficiency in resource-constrained environments.

In conclusion, each the models, Reka Flash and Reka Edge, introduced by Reka, successfully perform on LLM benchmarks using much smaller resources. It gives tight competition to existing state-of-the-art LLMs like Google’s Gemini Pro and OpenAI’s Gpt-4. Reka Flash uses knowledge in techniques like instruction tuning and reinforcement learning to excel in chat interactions. Meanwhile, Reka Edge stands out for its efficiency in local deployments.


Pragati Jhunjhunwala is a consulting intern at MarktechPost. She is currently pursuing her B.Tech from the Indian Institute of Technology(IIT), Kharagpur. She is a tech enthusiast and has a keen interest within the scope of software and data science applications. She is at all times reading concerning the developments in several field of AI and ML.


🚀 LLMWare Launches SLIMs: Small Specialized Function-Calling Models for Multi-Step Automation [Check out all the models]

LEAVE A REPLY

Please enter your comment!
Please enter your name here