MarkTechPost@AI 2024年12月03日
Liquid AI Introduces STAR: An AI Framework for the Automated Evolution of Tailored Architectures
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Liquid AI开发了STAR框架,旨在自动进化AI模型架构,以提高效率和性能。STAR通过基于线性输入变量系统(LIVs)理论创建新的架构搜索空间,重新构想模型构建过程。STAR使用进化优化原理进化架构设计,并通过迭代编译和评估基因组,实现架构的持续改进。STAR生成的架构在参数规模、效率和推理缓存方面优于手动调整的模型,例如Transformer和混合设计,为AI模型设计提供了新的方向。

🤔STAR框架通过基于线性输入变量系统(LIVs)理论创建新的架构搜索空间,重新定义了模型构建过程,并使用进化优化原理进化架构设计,从而实现模型架构的自动进化。

🧬STAR使用“STAR基因组”表示模型架构,这些基因组是架构设计的数值编码,包括骨干、算子和特征化基因组,共同决定模型中使用的计算单元的结构和属性。

🔄STAR通过评估、重组和突变等操作优化这些基因组,迭代地改进架构群体,从而生成多样化的架构设计,并优化质量、效率、大小和推理缓存等指标。

📈STAR生成的架构在参数数量、效率和推理缓存方面优于手动调整的模型,例如Transformer和混合设计,例如在保持或提高整体质量的同时,参数数量减少了13%。

💡STAR能够识别进化过程中出现的重复架构主题,为理解特定架构性能的原因提供宝贵的见解,从而推动未来AI模型设计领域的创新。

The development of effective AI models is crucial in deep learning research, but finding optimal model architectures remains challenging and costly. Traditional manual and automated approaches often fail to expand design possibilities beyond basic architectures like Transformers or hybrids, and the high cost of exploring a comprehensive search space limits model improvement. Manual optimization demands significant expertise and resources, while automated methods are often restricted by narrow search spaces, hindering substantial progress across tasks. To address these challenges, Liquid AI’s latest research offers a practical solution.

To address these challenges, Liquid AI has developed STAR (Synthesis of Tailored Architectures), a framework aimed at automatically evolving model architectures to enhance efficiency and performance. STAR reimagines the model-building process by creating a novel search space for architectures based on the theory of linear input-varying systems (LIVs). Unlike traditional methods that iterate on a limited set of known patterns, STAR provides a new approach to representing model structures, enabling exploration at different hierarchical levels through what they term “STAR genomes.”

These genomes serve as a numerical encoding of architecture designs, which STAR evolves using principles from evolutionary optimization. By compiling and evaluating these genomes iteratively, STAR allows for recombination and mutation, resulting in continuous refinements. The core idea is to treat model architectures as dynamic entities that can evolve over generations, optimizing for metrics like quality, efficiency, size, and inference cache—all key components of modern AI applications.

Technical Insights: STAR’s Architecture and Benefits

The technical foundation of STAR lies in its representation of model architectures as hierarchical numeric sequences—”genomes”—that define computational units and their interconnections. The search space is inspired by LIV systems, which generalize many common components of deep learning architectures, such as convolutional layers, attention mechanisms, and recurrent units. The STAR genome is composed of several levels of abstraction, including the backbone, operator, and featurizer genomes, which together determine the structure and properties of the computational units used in a model.

STAR optimizes these genomes through a combination of evolutionary algorithms. The process involves a series of operations: assessment, recombination, and mutation, which refine the population of architectures over time. Each architecture in the population is evaluated based on its performance on specific metrics, and the best-performing ones are recombined and mutated to form a new generation of architectures.

This approach enables STAR to generate diverse architectural designs. By breaking down architectures into manageable components and systematically optimizing them, STAR is capable of designing models that are efficient in terms of both computational requirements and quality. For instance, the STAR-generated architectures have shown improvements over manually tuned models such as Transformers and hybrid designs, especially when evaluated on parameters like size, efficiency, and inference cache requirements.

The implications of STAR are notable, especially given the challenges of scaling AI models while balancing efficiency and quality. Liquid AI’s results show that when optimizing for both quality and parameter size, STAR-evolved architectures consistently outperformed Transformer++ and hybrid models on downstream benchmarks. Specifically, STAR achieved a 13% reduction in parameter counts while maintaining or improving overall quality, measured by perplexity, across a variety of metrics and tasks.

The reduction in cache size is another important feature of STAR’s capabilities. When optimizing for quality and inference cache size, STAR-evolved models were found to have cache sizes up to 90% smaller than those of Transformer architectures while matching or surpassing them in quality. These improvements suggest that STAR’s approach of using evolutionary algorithms to synthesize architecture designs is viable and effective, particularly when optimizing for multiple metrics simultaneously.

Furthermore, STAR’s ability to identify recurring architecture motifs—patterns that emerge during the evolution process—provides valuable insights into the design principles that underlie the improvements observed. This analytical capability could serve as a tool for researchers looking to understand why certain architectures perform better, ultimately driving future innovation in AI model design.

Conclusion

STAR represents an important advancement in how we approach designing AI architectures. By leveraging evolutionary principles and a well-defined search space, Liquid AI has created a tool that can automatically generate tailored architectures optimized for specific needs. This framework is particularly valuable for addressing the need for efficient yet high-quality models capable of handling the diverse demands of real-world AI applications. As AI systems continue to grow in complexity, STAR’s approach offers a promising path forward—one that combines automation, adaptability, and insight to push the boundaries of AI model design.


Check out the Paper and Details. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter.. Don’t Forget to join our 60k+ ML SubReddit.

Evaluation of Large Language Model Vulnerabilities: A Comparative Analysis of Red Teaming Techniques’ Read the Full Report (Promoted)

The post Liquid AI Introduces STAR: An AI Framework for the Automated Evolution of Tailored Architectures appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI架构 自动进化 STAR Liquid AI 模型优化
相关文章