MarkTechPost@AI 2024年08月24日
AI21 Labs Released Jamba 1.5 Family of Open Models: Jamba 1.5 Mini and Jamba 1.5 Large Redefining Long-Context AI with Unmatched Speed, Quality, and Multilingual Capabilities for Global Enterprises
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

AI21 Labs发布Jamba 1.5系列开放模型,包括Jamba 1.5 Mini和Jamba 1.5 Large,该系列模型在处理长上下文任务等方面有突破,具有多种优势。

💪Jamba 1.5系列模型采用新颖的SSM-Transformer架构,能处理超长上下文,有效上下文窗口达256K令牌,对需分析和总结长文档的企业应用至关重要。

🚀在速度方面,Jamba 1.5模型在长上下文处理上比竞争对手快2.5倍,且在其规模类别内的所有上下文长度上都保持卓越性能,对需要快速处理任务的企业很关键。

🌟Jamba 1.5模型质量出众,Jamba 1.5 Mini在其规模类别中是最强的开放模型,Jamba 1.5 Large更是超越了领先模型,在不同基准测试中表现出色,能提供可靠准确的结果。

🌐Jamba 1.5模型具有多语言支持,能满足全球企业在多种语言环境下的需求,还为开发者提供了多种功能的原生支持,使其能适应各种开发需求。

🛠️Jamba 1.5模型可在多个平台上访问和部署,资源效率高,采用混合架构和量化技术,降低了内存占用,能在单个GPU上处理长上下文。

AI21 Labs has made a significant stride in the AI landscape by releasing the Jamba 1.5 family of open models, comprising Jamba 1.5 Mini and Jamba 1.5 Large. These models, built on the novel SSM-Transformer architecture, represent a breakthrough in AI technology, particularly in handling long-context tasks. AI21 Labs aims to democratize access to these powerful models by releasing them under the Jamba Open Model License, encouraging widespread experimentation and innovation.

Key Features of the Jamba 1.5 Models

One of the standout features of the Jamba 1.5 models is their ability to handle exceptionally long contexts. They boast an effective context window of 256K tokens, the longest in the market for open models. This feature is critical for enterprise applications requiring the analysis and summarization of lengthy documents. The models also excel in agentic and Retrieval-Augmented Generation (RAG) workflows, enhancing both the quality and efficiency of these processes.

Regarding speed, the Jamba 1.5 models are up to 2.5 times faster on long contexts than their competitors, and they maintain superior performance across all context lengths within their size class. This speed advantage is crucial for enterprises that need rapid turnaround times for tasks such as customer support or large-scale data processing.

The quality of the Jamba 1.5 models is another area where they outshine their peers. Jamba 1.5 Mini has been recognized as the strongest open model in its size class, achieving a score of 46.1 on the Arena Hard benchmark, outperforming larger models like Mixtral 8x22B and Command-R+. Jamba 1.5 Large goes even further, scoring 65.4, which surpasses leading models such as Llama 3.1 70B and 405B. This high-quality performance across different benchmarks highlights the robustness of the Jamba 1.5 models in delivering reliable and accurate results.

Multilingual Support and Developer Readiness

In addition to their technical prowess, the Jamba 1.5 models are designed with multilingual support, catering to languages such as Spanish, French, Portuguese, Italian, Dutch, German, Arabic, and Hebrew. This makes them versatile tools for global enterprises operating in diverse linguistic environments.

For developers, Jamba 1.5 models offer native support for structured JSON output, function calling, document object digestion, and citation generation. These features make the models adaptable to various development needs, enabling seamless integration into existing workflows.

Deployment and Efficiency

AI21 Labs has ensured that the Jamba 1.5 models are accessible and deployable across multiple platforms. They are available for immediate download on Hugging Face and are supported by major cloud providers, including Google Cloud Vertex AI, Microsoft Azure, and NVIDIA NIM. The models are expected to be available soon on additional platforms such as Amazon Bedrock, Databricks Marketplace, Snowflake Cortex, and others, making them easily deployable in various environments, including on-premises and virtual private clouds.

Another critical advantage of the Jamba 1.5 models is their resource efficiency. Built on a hybrid architecture that combines the strengths of Transformer and Mamba architectures, these models offer a lower memory footprint, allowing enterprises to handle extensive context lengths on a single GPU. AI21 Labs’ novel quantization technique, ExpertsInt8, further enhances this efficiency, which optimizes model performance without compromising quality.

Conclusion

The release of the Jamba 1.5 family by AI21 Labs marks a significant advancement in long-context handling. These models set new benchmarks in speed, quality, and efficiency and democratize access to cutting-edge AI technology through their open model license. As enterprises continue to seek AI solutions that deliver real-world value, the Jamba 1.5 models stand out as powerful tools capable of meeting the demands of complex, large-scale applications. Their availability across multiple platforms and support for multilingual environments further enhance their appeal, making them a versatile choice for developers and businesses.


Check out the Jamba 1.5 mini, Jamba 1.5 large, and Details. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter..

Don’t Forget to join our 49k+ ML SubReddit

Find Upcoming AI Webinars here

The post AI21 Labs Released Jamba 1.5 Family of Open Models: Jamba 1.5 Mini and Jamba 1.5 Large Redefining Long-Context AI with Unmatched Speed, Quality, and Multilingual Capabilities for Global Enterprises appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Jamba 1.5 AI21 Labs 长上下文处理 多语言支持 资源效率
相关文章