MarkTechPost@AI 07月11日 15:15
Mistral AI Releases Devstral 2507 for Code-Centric Language Modeling
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Mistral AI 与 All Hands AI 合作推出了 Devstral 2507 系列,包含 Devstral Small 1.1 和 Devstral Medium 2507 两款模型,专为开发者设计,支持基于代理的代码推理、程序合成和结构化任务执行。这些模型在性能和成本上进行了优化,适用于开发者工具和代码自动化系统。Devstral Small 1.1 是一款开源模型,支持本地和嵌入式使用;Devstral Medium 2507 则提供更高的准确性,仅通过 Mistral API 提供。

🚀 **Devstral Small 1.1:** 基于 Mistral-Small-3.1,拥有约 240 亿参数,支持 128k 令牌上下文窗口。该模型针对结构化输出进行了微调,包括 XML 和函数调用格式,适用于程序导航、多步编辑和代码搜索等任务。它在 SWE-Bench Verified 基准测试中达到了 53.6% 的成绩,并提供 Apache 2.0 许可,可用于研究和商业用途。

💡 **Devstral Medium 2507:** 仅通过 Mistral API 提供,拥有与 Small 版本相同的 128k 令牌上下文长度,但性能更高。在 SWE-Bench Verified 测试中,它的得分为 61.6%,优于 Gemini 2.5 Pro 和 GPT-4.1 等商业模型。由于其在长上下文中的强大推理能力,适合处理大型代码库或跨文件依赖项的代码代理任务。

💰 **定价与部署:** Devstral Small 1.1 提供了多种部署方式,包括 GGUF 量化版本,可在本地运行,方便开发者使用。Mistral API 对 Small 模型的定价为每百万输入令牌 0.10 美元,每百万输出令牌 0.30 美元。Devstral Medium 2507 的 API 定价为每百万输入令牌 0.40 美元,每百万输出令牌 2.00 美元,企业用户还可进行微调。

🛠️ **应用场景:** Devstral Small 适用于本地开发、实验或集成到客户端开发者工具,而 Devstral Medium 适用于对准确性和可靠性有较高要求的生产服务。两者均支持与 OpenHands 等代码代理框架集成,可用于测试生成、重构和错误修复等自动化工作流程。

Mistral AI, in collaboration with All Hands AI, has released updated versions of its developer-focused large language models under the Devstral 2507 label. The release includes two models—Devstral Small 1.1 and Devstral Medium 2507—designed to support agent-based code reasoning, program synthesis, and structured task execution across large software repositories. These models are optimized for performance and cost, making them applicable for real-world use in developer tools and code automation systems.

Devstral Small 1.1: Open Model for Local and Embedded Use

Devstral Small 1.1 (also called devstral-small-2507) is based on the Mistral-Small-3.1 foundation model and contains approximately 24 billion parameters. It supports a 128k token context window, which allows it to handle multi-file code inputs and long prompts typical in software engineering workflows.

The model is fine-tuned specifically for structured outputs, including XML and function-calling formats. This makes it compatible with agent frameworks such as OpenHands and suitable for tasks like program navigation, multi-step edits, and code search. It is licensed under Apache 2.0 and available for both research and commercial use.

Source: https://mistral.ai/news/devstral-2507

Performance: SWE-Bench Results

Devstral Small 1.1 achieves 53.6% on the SWE-Bench Verified benchmark, which evaluates the model’s ability to generate correct patches for real GitHub issues. This represents a noticeable improvement over the previous version (1.0) and places it ahead of other openly available models of comparable size. The results were obtained using the OpenHands scaffold, which provides a standard test environment for evaluating code agents.

While not at the level of the largest proprietary models, this version offers a balance between size, inference cost, and reasoning performance that is practical for many coding tasks.

Deployment: Local Inference and Quantization

The model is released in multiple formats. Quantized versions in GGUF are available for use with llama.cpp, vLLM, and LM Studio. These formats make it possible to run inference locally on high-memory GPUs (e.g., RTX 4090) or Apple Silicon machines with 32GB RAM or more. This is beneficial for developers or teams that prefer to operate without dependency on hosted APIs.

Mistral also makes the model available via their inference API. The current pricing is $0.10 per million input tokens and $0.30 per million output tokens, the same as other models in the Mistral-Small line.

Source: https://mistral.ai/news/devstral-2507

Devstral Medium 2507: Higher Accuracy, API-Only

Devstral Medium 2507 is not open-sourced and is only available through the Mistral API or through enterprise deployment agreements. It offers the same 128k token context length as the Small version but with higher performance.

The model scores 61.6% on SWE-Bench Verified, outperforming several commercial models, including Gemini 2.5 Pro and GPT-4.1, in the same evaluation framework. Its stronger reasoning capacity over long contexts makes it a candidate for code agents that operate across large monorepos or repositories with cross-file dependencies.

API pricing is set at $0.40 per million input tokens and $2 per million output tokens. Fine-tuning is available for enterprise users via the Mistral platform.

Comparison and Use Case Fit

ModelSWE-Bench VerifiedOpen SourceInput CostOutput CostContext Length
Devstral Small 1.153.6%Yes$0.10/M$0.30/M128k tokens
Devstral Medium61.6%No$0.40/M$2.00/M128k tokens

Devstral Small is more suitable for local development, experimentation, or integrating into client-side developer tools where control and efficiency are important. In contrast, Devstral Medium provides stronger accuracy and consistency in structured code-editing tasks and is intended for production services that benefit from higher performance despite increased cost.

Integration with Tooling and Agents

Both models are designed to support integration with code agent frameworks such as OpenHands. The support for structured function calls and XML output formats allows them to be integrated into automated workflows for test generation, refactoring, and bug fixing. This compatibility makes it easier to connect Devstral models to IDE plugins, version control bots, and internal CI/CD pipelines.

For example, developers can use Devstral Small for prototyping local workflows, while Devstral Medium can be used in production services that apply patches or triage pull requests based on model suggestions.

Conclusion

The Devstral 2507 release reflects a targeted update to Mistral’s code-oriented LLM stack, offering users a clearer tradeoff between inference cost and task accuracy. Devstral Small provides an accessible, open model with sufficient performance for many use cases, while Devstral Medium caters to applications where correctness and reliability are critical.

The availability of both models under different deployment options makes them relevant across various stages of the software engineering workflow—from experimental agent development to deployment in commercial environments.


Check out the Technical detailsDevstral Small model weights at Hugging Face and Devstral Medium will also be available on Mistral Code for enterprise customers and on finetuning API. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter, and Youtube and don’t forget to join our 100k+ ML SubReddit and Subscribe to our Newsletter.

The post Mistral AI Releases Devstral 2507 for Code-Centric Language Modeling appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Mistral AI Devstral 2507 代码生成 LLM
相关文章