Nvidia Blog 21小时前
Run LLMs on AnythingLLM Faster With NVIDIA RTX AI PCs
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

AnythingLLM 是一款专为AI爱好者设计的桌面应用,它允许用户在本地运行LLMs,RAG系统和智能工具。通过支持NVIDIA NIM微服务,AnythingLLM在NVIDIA GeForce RTX和NVIDIA RTX PRO GPU上实现了更快的性能,提供了更流畅的AI工作流程。用户可以使用它进行问答、个人数据查询、文档摘要和数据分析等任务。AnythingLLM易于安装和使用,支持多种开源和云端LLMs,是AI爱好者,特别是配备GeForce RTX和NVIDIA RTX PRO GPU的用户的理想选择。

💡 AnythingLLM 是一款多功能的AI应用,用户可以在本地运行大型语言模型(LLMs)、检索增强生成(RAG)系统和智能工具。它充当用户首选LLMs和数据之间的桥梁,并支持访问各种工具(称为技能),从而更轻松高效地使用LLMs。

🚀 AnythingLLM 通过支持NVIDIA NIM微服务,在NVIDIA GeForce RTX和NVIDIA RTX PRO GPU上实现了加速。NVIDIA NIM微服务是性能优化的预打包生成式AI模型,通过简化的API,使用户能够在RTX AI PC上轻松启动AI工作流程。

⚙️ AnythingLLM 提供多种应用场景,包括问答、个人数据查询、文档摘要和数据分析。它可以连接到各种开源本地LLMs以及云端LLMs,如OpenAI、Microsoft和Anthropic。用户还可以通过社区中心访问技能,扩展其智能AI的功能。

Large language models (LLMs), trained on datasets with billions of tokens, can generate high-quality content. They’re the backbone for many of the most popular AI applications, including chatbots, assistants, code generators and much more.

One of today’s most accessible ways to work with LLMs is with AnythingLLM, a desktop app built for enthusiasts who want an all-in-one, privacy-focused AI assistant directly on their PC.

With new support for NVIDIA NIM microservices on NVIDIA GeForce RTX and NVIDIA RTX PRO GPUs, AnythingLLM users can now get even faster performance for more responsive local AI workflows.

What Is AnythingLLM?

AnythingLLM is an all-in-one AI application that lets users run local LLMs, retrieval-augmented generation (RAG) systems and agentic tools.

It acts as a bridge between a user’s preferred LLMs and their data, and enables access to tools (called skills), making it easier and more efficient to use LLMs for specific tasks like:

AnythingLLM can connect to a wide variety of open-source local LLMs, as well as larger LLMs in the cloud, including those provided by OpenAI, Microsoft and Anthropic. In addition, the application provides access to skills for extending its agentic AI capabilities via its community hub.

With a one-click install and the ability to launch as a standalone app or browser extension — wrapped in an intuitive experience with no complicated setup required — AnythingLLM is a great option for AI enthusiasts, especially those with GeForce RTX and NVIDIA RTX PRO GPU-equipped systems.

RTX Powers AnythingLLM Acceleration

GeForce RTX and NVIDIA RTX PRO GPUs offer significant performance gains for running LLMs and agents in AnythingLLM — speeding up inference with Tensor Cores designed to accelerate AI.

AnythingLLM runs LLMs with Ollama for on-device execution accelerated through Llama.cpp and ggml tensor libraries for machine learning.

Ollama, Llama.cpp and GGML are optimized for NVIDIA RTX GPUs and the fifth-generation Tensor Cores. Performance on GeForce RTX 5090 is 2.4X compared to an Apple M3 Ultra.

GeForce RTX 5090 delivers 2.4x faster LLM inference in AnythingLLM than Apple M3 Ultra on both Llama 3.1 8B and DeepSeek R1 8B.

As NVIDIA adds new NIM microservices and reference workflows — like its growing library of AI Blueprints — tools like AnythingLLM will unlock even more multimodal AI use cases.

AnythingLLM — Now With NVIDIA NIM

AnythingLLM recently added support for NVIDIA NIM microservices — performance-optimized, prepackaged generative AI models that make it easy to get started with AI workflows on RTX AI PCs with a streamlined API.

NVIDIA NIMs are great for developers looking for a quick way to test a Generative AI model in a workflow. Instead of having to find the right model, download all the files and figure out how to connect everything, they provide a single container that has everything you need. And they can run both on Cloud and PC, making it easy to prototype locally and then deploy on the cloud.

By offering them within AnythingLLM’s user-friendly UI, users have a quick way to test them and experiment with them. And then they can either connect them to their workflows with AnythingLLM, or leverage NVIDIA AI Blueprints and NIM documentation and sample code to plug them directly to their apps or projects.

Explore the wide variety of NIM microservices available to elevate AI-powered workflows, including language and image generation, computer vision and speech processing.

Each week, the RTX AI Garage blog series features community-driven AI innovations and content for those looking to learn more about NIM microservices and AI Blueprints, as well as building AI agents, creative workflows, digital humans, productivity apps and more on AI PCs and workstations. 

Plug in to NVIDIA AI PC on Facebook, Instagram, TikTok and X — and stay informed by subscribing to the RTX AI PC newsletter.

Follow NVIDIA Workstation on LinkedIn and XSee notice regarding software product information.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AnythingLLM LLMs NVIDIA NIM AI加速
相关文章