MarkTechPost@AI 2024年10月02日
Instructive Decoding (ID): A Novel AI Method that Enhances the Attention of Instruction-Tuned LLMs Towards Provided Instructions during the Generation Phase without Any Parameter Updates
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Instructive Decoding(ID)是一种无需参数更新,能增强指令调优语言模型的方法。该方法利用“噪声指令”创造对比方法,改善模型在不同任务中的表现,提高对指令的遵循度和整体响应质量,实验证明其在多种模型和任务中具有有效性。

🧐Instructive Decoding(ID)是KAIST AI的研究人员提出的方法,通过使用“噪声指令”,即原始指令的变体,以对比预测下一个标记,引导模型输出向不同方向发展,尤其是使用“相反”指令来提升模型性能。

📈实验表明,ID在准确性上有显著提升,较小的模型经ID增强后性能可超过较大的模型,该方法提高了对指令的遵循度,增强了整体响应质量,在各种模型和任务中都显示出了有效性。

🎯Instruction-tuning可使预训练语言模型更好地遵循自然语言指令,提高对未见任务的泛化能力,尤其在零样本场景中。ID在此基础上,利用噪声指令来提升指令调优语言模型的泛化能力。

🔬实验使用SUPNATINST和UNNATINST数据集,评估了Tk-Instruct、Alpaca和T0等模型在语法错误纠正和文本蕴含等任务上的表现,通过Rouge-L、Exact Match(EM)、Label Adherence(LA)和Label Coherence(LC)等指标评估性能,ID持续改进了结果。

Instruction-tuned LMs have shown remarkable zero-shot generalization but often fail on tasks outside their training data. These LMs, built on large datasets and billions of parameters, excel in In-Context Learning (ICL), generating responses based on a few examples without re-training. However, the training dataset’s scope limits its effectiveness on unfamiliar tasks. Techniques like prompt engineering and output diversification help improve performance but require significant effort. Recent research explores applying the cognitive anchoring effect to LMs, suggesting that emphasizing initial prompts can enhance task-specific responses and improve fidelity to instructions.

Researchers from KAIST AI introduced Instructive Decoding (ID), a method that enhances instruction-tuned LMs without parameter updates. ID uses “noisy instructions,” altered versions of the original instructions, to create a contrastive approach for predicting the next token. By steering the model’s output in different directions, especially using “opposite” instructions, ID improves model performance across tasks. Experiments show significant gains in accuracy, with smaller models enhanced by ID outperforming larger ones. This method improves adherence to instructions and enhances overall response quality, demonstrating its effectiveness across various models and tasks.

Instruction-tuning fine-tunes pre-trained LMs to follow natural language instructions better, improving generalization to unseen tasks, especially in zero-shot scenarios. Expanding the variety and complexity of training tasks enhances this capability, although the models often rely heavily on pre-trained knowledge. Prior research highlights that LMs are sensitive to familiar instructions, even handling misleading ones, and this sensitivity can be leveraged through contrastive techniques. Contrast in text generation, like Contrastive Decoding, compares outputs from different models or inputs to improve performance. This study extends these ideas by using noisy instructions to boost generalization in instruction-tuned LMs.

Instructive Decoding improves response generation in instruction-tuned models by contrasting outputs generated from noisy instructions. It builds on the anchoring effect, where initial information influences subsequent judgments and leverages differences between responses generated from original and altered instructions. The method uses noisy instruction variants like truncated, shuffled, or random words to mislead the model while ensuring task fidelity. By comparing logits from original and noisy instructions during decoding, Instructive Decoding helps models correct biases and produce responses more aligned with the intended instructions, refining their performance on unseen tasks.

The experimental setup uses the SUPNATINST and UNNATINST datasets, evaluating models like Tk-Instruct, Alpaca, and T0 across tasks like Grammar Error Correction and Textual Entailment. Rouge-L, Exact Match (EM), Label Adherence (LA), and Label Coherence (LC) metrics assess performance. ID consistently improves results, especially for larger models like Tk-XXL, enhancing LA and LC. Interestingly, noisy instructions enhance output quality with ID despite baseline performance degradation. Though task-specific performance varies, the ‘opposite’ instruction variant proves robust across tasks. Overall, ID shows significant gains across model sizes and task types.

The study investigates the challenges of unseen task generalization in instruction-tuned language models. The proposed method, ID, leverages the anchoring effect using “noisy” instructions to counteract inherent model biases. By contrasting predictions with those generated from altered instructions, ID enhances model performance, particularly with the “opposite” noisy variant, which deviates most from the original input. Empirical results show ID’s effectiveness across multiple tasks, with notable improvements in prediction diversity. The approach requires no additional parameter updates, making it a practical tool for improving instruction-following in language models.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter..

Don’t Forget to join our 50k+ ML SubReddit

The post Instructive Decoding (ID): A Novel AI Method that Enhances the Attention of Instruction-Tuned LLMs Towards Provided Instructions during the Generation Phase without Any Parameter Updates appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Instructive Decoding 语言模型 指令调优 模型性能
相关文章