EnterpriseAI 2024年10月03日
Liquid AI’s LFM Models Challenge the Status Quo, Outperforming Industry Giants
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Liquid AI,一家由麻省理工学院计算机科学与人工智能实验室 (CSAIL) 前研究人员共同创办的初创公司,推出了一系列新型 Liquid Foundation Models (LFMs),旨在改变人工智能领域。这些模型采用了一种突破性的架构,与目前大多数 GenAI 系统(包括 OpenAI 的 ChatGPT)使用的 Transformer 模型截然不同。Liquid 从“第一性原理”构建基础模型,这意味着这些模型从头开始构建,不依赖于 Transformer 等现有框架。Liquid 声称其 LFMs 具有卓越的性能,在某些方面甚至超越了同等规模的领先大型语言模型。LFMs 的核心优势在于它们能够在消耗明显更少内存的情况下超越基于 Transformer 的模型。LFMs 的运营效率使其成为跨行业广泛用例的理想选择。与传统的 LLM 在长上下文处理期间内存使用量大幅增加不同,LFMs 保持着更小的内存占用量。这种效率使它们特别适合需要处理大量顺序数据的应用程序,例如文档分析和 AI 聊天机器人。

🤔 Liquid AI 推出的 Liquid Foundation Models (LFMs) 采用了一种全新的架构,摆脱了传统 Transformer 模型的束缚,从“第一性原理”构建基础模型,不依赖于现有框架。

💪 Liquid AI 声称其 LFMs 具有卓越的性能,在某些方面甚至超越了同等规模的领先大型语言模型。例如,LFM 1.3B 在多个第三方基准测试中表现优于 Meta 的 Llama 3.1-8B 和微软的 Phi-3.5 3.8B,包括广受认可的大规模多任务语言理解 (MMLU) 基准测试。

🧠 LFMs 的核心优势在于它们能够在消耗明显更少内存的情况下超越基于 Transformer 的模型。LFMs 的运营效率使其成为跨行业广泛用例的理想选择,特别适合需要处理大量顺序数据的应用程序,例如文档分析和 AI 聊天机器人。

🔓 目前,Liquid AI 的 LFMs 尚未开源,用户只能通过 Liquid 的推理游乐场、Lambda Chat UI 和 API 以及 Perplexity AI 访问。Liquid 计划很快在 Cerebras Inference 上提供访问权限,并且正在优化 LFM 堆栈以适用于 NVIDIA、AMD、Qualcomm、Cerebras 和 Apple 硬件。

🚀 Liquid AI 计划发布一系列技术博客文章,深入探讨每个模型的内部工作原理,并通过科学和技术报告公开发布其研究成果和方法,以支持用户和 AI 社区,并推进 AI 领域的发展。

Liquid AI, a startup co-founded by former researchers from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL), has introduced a new series of Liquid Foundation Models (LFMs) that aim to transform the AI landscape.

What sets these models apart is their groundbreaking architecture, which breaks away from the transformer model used in most of today's GenAI systems, including OpenAI’s ChatGPT. The transformer model was introduced in the seminal paper "Attention is All You Need" by Vaswani et al. in 2017. 

Liquid has taken a different approach by building foundation models on “first principles”, meaning these models are built from the ground up, without relying on existing frameworks like transformers. 

"Architecture work cannot happen in a vacuum – our goal is to develop useful models that are competitive with the current best-in-class LLMs," noted Liquid via their website. "In doing so, we hope to show that model performance isn’t just about scale – it’s also about innovation."

The startup’s LFMs are available in three variants: LFM 1.3B, LFM 3B, and LFM 40 MoE. "MoE" stands for “Mixture of Experts,” while the "B" indicates the number of parameters in billions. Typically, models with more parameters offer better performance.

(Maxisport/Shutterstock)

The Boston-based startup claims that its LFMs offer exceptional performance, matching or even surpassing some of the leading large language models of comparable size. 

The LFM 1.3B demonstrated superior performance compared to Meta’s Llama 3.1-8B and Microsoft’s Phi-3.5 3.8B across several third-party benchmarks, including the widely recognized Massive Multitask Language Understanding (MMLU). Liquid claims that this marks “the first instance in which a non-GPT architecture has notably surpassed transformer-based models.”

The core advantage of LFMs is their ability to outperform transformer-based models while consuming significantly less memory. The operational efficiency of LFMs makes them ideal for a wide range of use cases across industries. 

In contrast to traditional LLMs, which encounter a substantial increase in memory usage during long-context processing, LFMs maintain a much smaller memory footprint. This efficiency makes them particularly well-suited for applications that require the processing of large volumes of sequential data, such as document analysis and AI chatbots.

It’s important for users to note that Liquid’s LFMs are not open source. Instead, access is available exclusively through Liquid’s inference playground, Lambda Chat UI and API, or Perplexity AI. Liquid plans on making it available on Cerebars Inference soon, and is also optimizing the LFM stack for NVIDIA, AMD, Qualcomm, Cerebras, and Apple hardware.

To further support users and the AI community, Liquid AI plans to publish a series of technical blog posts that delve into the inner workings of each model

In a statement from Liquid AI, the company emphasized, “At Liquid AI, we take an open-science approach. We have and will continue to contribute to the advancement of the AI field by openly publishing our findings and methods through scientific and technical reports. As part of this commitment, we will release relevant data and models produced by our research efforts to the wider AI community.”

(SomYuZu/Shutterstock)

“We have dedicated a lot of time and resources to developing these architectures, so we're not open-sourcing our models at the moment. This allows us to continue building on our progress and maintain our edge in the competitive AI landscape.”

The four founders of Liquid - Daniela Rus, Mathias Lechner, Alexander Amini, and Ramin Hasani are recognized as pioneers in the concept of “liquid neural networks.” This innovative approach allows AI models to adapt dynamically to new data and tasks.

Liquid was created with the mission to transcend generative pre-trained transformers (GPT) and develop capable, efficient general-purpose AI systems that can operate effectively at any scale. The startup emerged from stealth late last year, securing $37.5 million in a two-stage seed round, with a reported valuation of $303 million.

Liquid’s AI models exemplify the rapid innovation taking place in the highly competitive AI market. Newcomers like Liquid are making significant strides; however, the industry's evolving landscape means that even more advanced models will continue to appear. This ongoing progress will likely foster diversity in model architectures, contributing to a more balanced competitive environment with a more level playing field. 

 

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Liquid AI Liquid Foundation Models LFMs Transformer 人工智能
相关文章