MarkTechPost@AI 2024年11月30日
PRIME Intellect Releases INTELLECT-1 (Instruct + Base): The First 10B Parameter Language Model Collaboratively Trained Across the Globe
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

PRIME Intellect发布了名为INTELLECT-1的100亿参数语言模型,这是全球首个通过分布式协作训练的LLM。该模型利用PRIME框架,克服了分布式训练中网络不稳定和计算节点动态变化等挑战,在三个大陆上利用了多达112个H100 GPU,并在最佳条件下实现了高达96%的计算利用率。INTELLECT-1的训练使用了1万亿个token,涵盖了各种领域,并通过创新技术确保了训练的稳定性和容错性。该项目标志着LLM训练向更广泛的群体开放,为社区驱动的AI项目树立了先例,并有望促进AI领域的公平与包容性。

🤔 **全球首个协同训练的100亿参数语言模型:**INTELLECT-1由PRIME Intellect发布,是首个通过分布式协作方式训练的100亿参数语言模型,证明了利用全球分散资源训练先进LLM的可行性,打破了大型科技公司对先进AI技术的垄断。

💻 **PRIME框架克服分布式训练挑战:**PRIME框架旨在解决分布式训练中网络可靠性和计算节点动态变化等问题,实现了高达96%的计算利用率,并通过弹性设备网格、混合训练方法和8位量化策略等技术提升了训练效率和容错性。

📚 **基于海量数据训练:**INTELLECT-1的训练使用了1万亿个token,涵盖了各种领域的数据,包括公开数据集和PRIME Intellect及其合作伙伴整理的专有数据集,确保了模型对不同领域的广泛理解。

📊 **性能表现与未来展望:**INTELLECT-1在多个基准测试中表现出色,例如MMLU和HellaSwag,虽然略逊于一些顶尖的集中式训练模型,但其在分布式训练环境下的表现已经非常出色。随着模型规模的扩大和通信策略的改进,集中式和分布式训练之间的差距将进一步缩小。

🌍 **社区协作推动AI发展:**INTELLECT-1的训练由全球30个独立计算贡献者参与,彰显了社区驱动的AI项目的可扩展性,为未来更广泛的AI研究合作奠定了基础,也为AI领域带来了更多可能性,让AI变得更加开放和包容。

In recent years, the evolution of artificial intelligence has brought forth increasingly sophisticated large language models (LLMs). However, training these models remains a complex challenge due to their immense computational requirements. Traditionally, training such models has been possible only in centralized environments with high-bandwidth interconnects, typically within large data centers controlled by a few tech giants. This centralized paradigm limits accessibility, as it requires significant resources that only a few organizations can afford. These restrictions have raised concerns about equitable access to advanced AI technologies and their potential monopolization. To address these barriers, researchers have begun exploring collaborative, decentralized training approaches. The challenge lies in overcoming issues such as low inter-node bandwidth and unpredictable node availability, which make decentralized training more complex than its centralized counterpart.

The Release of INTELLECT-1

PRIME Intellect has released INTELLECT-1 (Instruct + Base), the first 10-billion-parameter language model collaboratively trained across the globe. This model demonstrates the feasibility of using decentralized, community-driven resources for training advanced LLMs. PRIME Intellect utilized their PRIME framework, specifically designed to overcome the challenges of decentralized training, including network unreliability and the dynamic addition or removal of compute nodes. The framework utilized up to 112 H100 GPUs across three continents and achieved a compute utilization rate of up to 96% under optimal conditions, demonstrating that decentralized training can match the performance levels of traditional setups. This approach broadens access to high-performance AI models and fosters a collaborative research environment where contributors worldwide can participate in AI development.

Technical Details

According to the official release, INTELLECT-1 was developed using a diverse mix of high-quality datasets, including publicly available data and proprietary datasets curated by PRIME Intellect and their partners. The model was trained on 1 trillion tokens, ensuring it has a broad understanding of various domains. The training process involved 14 concurrent nodes distributed across three continents, with compute sponsors dynamically joining and leaving as needed. This dynamic approach allowed for significant flexibility, which is crucial for real-world deployment scenarios. PRIME Intellect also ensured training stability through innovations like live checkpointing and fault-tolerant communication, enabled by the PRIME framework.

Technically, INTELLECT-1’s training was made possible through innovations in the PRIME framework, which addressed the constraints of geographically distributed nodes. PRIME features the ElasticDeviceMesh, an abstraction that manages both internet-wide communication and local, fault-tolerant data-sharing across nodes. Hybrid training approaches combining Fully Sharded Data Parallel (FSDP) techniques for intra-node efficiency and Distributed Low-Communication (DiLoCo) algorithms for minimal inter-node communication were implemented. To minimize bandwidth requirements, the PRIME framework included an 8-bit quantization strategy for gradient transfers, reducing the communication payload by up to 400 times compared to traditional data-parallel training. Fault tolerance was managed through dynamic node management, allowing new nodes to join seamlessly and failed nodes to be removed with minimal disruption. These innovations facilitated effective decentralized model training while maintaining high computational efficiency.

Benchmark Results and Implications

The release of INTELLECT-1 marks a significant step forward in making LLM training accessible beyond large corporations. Results from the training process reveal a model that competes with similarly sized models trained in centralized settings. For instance, INTELLECT-1 achieved 37.5% accuracy on the MMLU benchmark and 72.26% on HellaSwag. Additionally, INTELLECT-1 outperformed several other open-source models in specific benchmarks, including 65.82% on the WinoGrande challenge. Although these figures slightly lag behind some state-of-the-art centralized models, the results are notable given the challenges of decentralized training. More importantly, this experiment sets a precedent for large-scale collaborations and paves the way for further developments in community-led AI projects. The global network of 30 independent compute contributors not only ensured the success of the project but also highlighted the scalability of such efforts. As decentralized models grow in scale and as communication strategies improve, the gap between centralized and decentralized training will likely continue to close.

Conclusion

The release of INTELLECT-1 represents a milestone in the pursuit of more accessible AI research. By leveraging decentralized resources to train a 10-billion-parameter language model, PRIME Intellect and its collaborators have demonstrated that advanced AI development need not be limited to a few elite corporations. Through innovations in distributed training frameworks and global collaboration, INTELLECT-1 sets a new standard for what is possible in open and inclusive AI research. The PRIME framework, along with the publicly available INTELLECT-1 model and training data, will hopefully inspire more community-driven projects, helping to level the playing field in the AI space and opening doors for more diverse contributions. This is an important step towards making AI an accessible and inclusive resource for everyone.


Check out the Paper, Details, and Models on Hugging Face (Instruct and Base). All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter.. Don’t Forget to join our 59k+ ML SubReddit.

Evaluation of Large Language Model Vulnerabilities: A Comparative Analysis of Red Teaming Techniques’ Read the Full Report (Promoted)

The post PRIME Intellect Releases INTELLECT-1 (Instruct + Base): The First 10B Parameter Language Model Collaboratively Trained Across the Globe appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

INTELLECT-1 分布式训练 语言模型 PRIME框架 人工智能
相关文章