MarkTechPost@AI 2024年12月07日
Ruliad AI Releases DeepThought-8B: A New Small Language Model Built on LLaMA-3.1 with Test-Time Compute Scaling and Deliverers Transparent Reasoning
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Ruliad AI发布了名为Deepthought-8B的新型语言模型,它基于LLaMA-3.1架构,拥有80亿参数。该模型专注于推理透明度和可控性,旨在提供与更大模型相当的复杂问题解决能力,同时保持高效运行。Deepthought-8B的特点包括透明的推理机制,以结构化的JSON格式输出推理步骤,并支持可编程的推理模式,无需重新训练即可适应不同任务。此外,它还具有测试时间计算可扩展性,能够根据任务复杂度调整推理深度。该模型在编码和数学任务等方面表现出色,但也存在一些局限性,例如复杂数学推理、长上下文处理和边缘情况处理等。

🤔 **透明的推理机制:**Deepthought-8B 的核心特点是其透明的推理机制,它会将决策过程中的每个步骤都记录下来,并以结构化的 JSON 格式输出。用户可以清晰地追踪模型的思考过程,从而增强对模型输出结果的信任,并方便将其集成到需要清晰可解释的 AI 逻辑的应用中。

⚙️ **可编程的推理模式:**与许多需要针对不同任务进行重新训练的模型不同,Deepthought-8B 允许用户自定义推理方法,而无需进行重新训练。这种适应性使其适用于各种应用场景,从编码任务到复杂问题解决。

📈 **测试时间计算可扩展性:**Deepthought-8B 能够根据任务的复杂度调整推理深度,从而在不同挑战中提供灵活的工具。例如,对于简单的任务,它可以快速给出答案,而对于复杂的推理任务,则会进行更深入的分析。

🖥️ **高效的运行环境:**Deepthought-8B 可以在拥有 16GB 或更多 VRAM 的系统上高效运行,并支持 Flash Attention 2 等高级功能以提高性能。其技术生态系统建立在 Python、PyTorch 和 Transformers 库等广泛使用的框架之上,方便开发者使用。

⚠️ **模型局限性:**Deepthought-8B 在复杂数学推理、长上下文处理和边缘情况处理等方面仍有提升空间。Ruliad AI 对模型的局限性进行了公开说明,体现了其透明度,并鼓励用户提供反馈以改进未来版本。

Ruliad AI released Deepthought-8B-LLaMA-v0.01-alpha, focusing on reasoning transparency and control. This model, built on LLaMA-3.1 with 8 billion parameters, is designed to offer sophisticated problem-solving capabilities comparable to much larger models while maintaining operational efficiency.

Deepthought-8B distinguishes itself with unique features aimed at making AI reasoning more accessible and understandable. The standout characteristic is its transparent reasoning mechanism, where every step in the decision-making process is documented. This feature ensures users can follow the model’s thought process, outputted in a structured JSON format. This step-by-step reasoning builds trust in its outputs and facilitates seamless integration into applications requiring clear and explainable AI logic. Another aspect of Deepthought-8B is its programmable reasoning patterns. Unlike many models that require retraining for different tasks, this model allows customization of reasoning approaches without necessitating retraining. This adaptability makes it suitable for various applications, from coding tasks to complex problem-solving scenarios. Also, its scalability in test-time computing ensures it can adjust reasoning depth based on the complexity of tasks, providing users with a versatile tool for various challenges.

Deepthought-8B operates efficiently on systems with 16GB or more VRAM and supports advanced features like Flash Attention 2 for enhanced performance. Its technical ecosystem is built on widely used frameworks such as Python, PyTorch, and the Transformers library, allowing developers compatibility and ease of use. Each reasoning chain in the model includes stages such as problem understanding, data gathering, analysis, calculation, verification, conclusion drawing, and implementation. These clearly defined steps enhance the model’s usability and position it as a valuable tool for domains requiring rigorous logical workflows.

Deepthought-8B also shows strong performance across various benchmarks, like coding and mathematical tasks effectively. However, it has limitations. Complex mathematical reasoning, long-context processing, and edge-case handling are areas where the model has room for improvement. Acknowledging these limitations reflects Ruliad’s transparency in presenting the model’s capabilities, fostering user trust, and encouraging constructive feedback for future iterations. Ruliad has positioned Deepthought-8B as a commercial enterprise solution, with licensing terms supporting this approach. The model is accompanied by comprehensive support options, including social media and email contact, ensuring users can easily access assistance. The documentation for Deepthought-8B includes detailed installation and usage guidelines.

Installation

pip install torch transformers# Optional: Install Flash Attention 2 for better performancepip install flash-attn

Usage

1.First, set your HuggingFace token as an environment variable:

      export HF_TOKEN=your_token_hereexport HF_HUB_ENABLE_HF_TRANSFER=1

      2.Use the model in your Python code:

        from transformers import AutoModelForCausalLM, AutoTokenizerimport torch# Initialize the modelmodel_name = "ruliad/deepthought-8b-llama-v0.01-alpha"tokenizer = AutoTokenizer.from_pretrained(    model_name,    add_bos_token=False,    trust_remote_code=True,    padding="left",    torch_dtype=torch.bfloat16,)model = AutoModelForCausalLM.from_pretrained(    model_name,    torch_dtype=torch.bfloat16,    device_map="auto",    attn_implementation="flash_attention_2",  # Use "eager" (or omit) if flash_attn is not installed    use_cache=True,    trust_remote_code=True,)

        3.Run the provided example script:

        python deepthought_inference.py

          In conclusion, Deepthought-8B, with its 8.03 billion parameters, rivals larger 70B-scale models in reasoning tasks, leveraging advanced features such as JSON-structured outputs and customizable inference paths. Its ability to run on systems with as little as 16GB VRAM ensures accessibility, while test-time compute scaling allows users to tailor performance to task complexity. With over 10,000 downloads in the past month, the model’s adoption underscores its utility and relevance.


          Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter.. Don’t Forget to join our 60k+ ML SubReddit.

          [Partner with us]: ‘Next Magazine/Report- Open Source AI in Production’

          The post Ruliad AI Releases DeepThought-8B: A New Small Language Model Built on LLaMA-3.1 with Test-Time Compute Scaling and Deliverers Transparent Reasoning appeared first on MarkTechPost.

          Fish AI Reader

          Fish AI Reader

          AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

          FishAI

          FishAI

          鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

          联系邮箱 441953276@qq.com

          相关标签

          Deepthought-8B LLaMA 推理模型 透明推理 可编程推理
          相关文章