MarkTechPost@AI 03月12日
Google AI Releases Gemma 3: Lightweight Multimodal Open Models for Efficient and On‑Device AI
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Google DeepMind推出了Gemma 3系列开放模型,旨在解决AI领域中计算资源需求高和部署困难的问题。Gemma 3基于Gemini 2.0技术,可在单个GPU或TPU上高效运行,提供1B、4B、12B和27B等多种尺寸选择,支持预训练和指令调优。这些模型具备多模态和多语言能力,能够处理文本和图像,并支持超过140种语言。Gemma 3还拥有更大的上下文窗口,适用于处理大量信息的任务。通过强化学习和安全措施,Gemma 3致力于实现高效、安全和广泛可访问的AI。

💡Gemma 3系列模型旨在解决AI模型对计算资源的高需求问题,并实现更广泛的应用,它可以在单个GPU或TPU上高效运行。

🌍Gemma 3具备多模态和多语言能力,4B、12B和27B模型能够处理文本和图像,并支持超过140种语言,适用于服务全球用户。

📚Gemma 3拥有更大的上下文窗口,128,000 tokens (1B模型为32,000 tokens),适合处理大量信息的任务,例如文档摘要和长对话管理。

🛡️Gemma 3的训练过程结合了人类反馈强化学习和其他后训练方法,旨在确保模型输出符合用户预期,并维护使用安全。

In the field of artificial intelligence, two persistent challenges remain. Many advanced language models require significant computational resources, which limits their use by smaller organizations and individual developers. Additionally, even when these models are available, their latency and size often make them unsuitable for deployment on everyday devices such as laptops or smartphones. There is also an ongoing need to ensure these models operate safely, with proper risk assessments and built‑in safeguards. These challenges have motivated the search for models that are both efficient and broadly accessible without compromising performance or security.

Google AI Releases Gemma 3: A Collection of Open Models

Google DeepMind has introduced Gemma 3—a family of open models designed to address these challenges. Developed with technology similar to that used for Gemini 2.0, Gemma 3 is intended to run efficiently on a single GPU or TPU. The models are available in various sizes—1B, 4B, 12B, and 27B—with options for both pre‑trained and instruction‑tuned variants. This range allows users to select the model that best fits their hardware and specific application needs, making it easier for a wider community to incorporate AI into their projects.

Technical Innovations and Key Benefits

Gemma 3 is built to offer practical advantages in several key areas:

Performance Insights and Evaluations

Early evaluations of Gemma 3 indicate that the models perform reliably within their size class. In one set of tests, the 27B variant achieved a score of 1338 on a relevant leaderboard, indicating its capacity to deliver consistent and high‐quality responses without requiring extensive hardware resources. Benchmarks also show that the models are effective at handling both text and visual data, thanks in part to a vision encoder that manages high-resolution images with an adaptive approach.

The training of these models involved a large and varied dataset of text and images—up to 14 trillion tokens for the largest variant. This comprehensive training regimen supports their ability to address a wide range of tasks, from language understanding to visual analysis. The widespread adoption of earlier Gemma models, along with a vibrant community that has already produced numerous variants, underscores the practical value and reliability of this approach.

Conclusion: A Thoughtful Approach to Open, Accessible AI

Gemma 3 represents a careful step toward making advanced AI more accessible. Available in four sizes and capable of processing both text and images in over 140 languages, these models offer an expanded context window and are optimized for efficiency on everyday hardware. Their design emphasizes a balanced approach—delivering robust performance while incorporating measures to ensure safe use.

In essence, Gemma 3 is a practical solution to longstanding challenges in AI deployment. It allows developers to integrate sophisticated language and vision capabilities into a variety of applications, all while maintaining an emphasis on accessibility, reliability, and responsible usage.


Check out the Models on Hugging Face and Technical details. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 80k+ ML SubReddit.

Meet Parlant: An LLM-first conversational AI framework designed to provide developers with the control and precision they need over their AI customer service agents, utilizing behavioral guidelines and runtime supervision. It’s operated using an easy-to-use CLI and native client SDKs in Python and TypeScript .

The post Google AI Releases Gemma 3: Lightweight Multimodal Open Models for Efficient and On‑Device AI appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Gemma 3 Google AI 开放模型 多模态 轻量级AI
相关文章