MarkTechPost@AI 02月02日
Can AI Understand Subtext? A New AI Approach to Natural Language Inference
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

当前自然语言推理(NLI)模型在识别隐式蕴含方面存在困难,因为现有数据集主要关注显式蕴含。为了解决这个问题,谷歌DeepMind和宾夕法尼亚大学的研究人员提出了Implied NLI(INLI)数据集。该数据集通过结构化的蕴含框架和Gemini-Pro的少样本提示方法,系统地引入了隐式含义,并生成了高质量的隐式蕴含。实验结果表明,在INLI上微调的模型在检测隐式蕴含方面取得了显著提升,并且在不同领域具有良好的泛化能力。这一研究为训练更精细、更具上下文感知能力的自然语言理解模型奠定了基础。

🧠 当前NLI模型难以识别隐式含义,主要原因是现有数据集侧重于显式蕴含,导致模型在处理间接表达的含义时表现不佳,限制了对话AI等应用的发展。

💡 研究人员提出了INLI数据集,通过将结构化蕴含框架转化为<前提,隐式蕴含>对,并结合显式蕴含、中性假设和矛盾,构建了一个全面的训练集。利用Gemini-Pro的少样本提示,生成高质量的隐式蕴含,同时降低标注成本。

🎯 在INLI上微调的模型在检测隐式蕴含方面取得了显著提升,准确率高达92.5%,远高于在传统NLI数据集上微调的模型。此外,该模型在未见过的数据集上表现出良好的泛化能力,证明了INLI的稳健性。

Understanding implicit meaning is a fundamental aspect of human communication. Yet, current Natural Language Inference (NLI) models struggle to recognize implied entailments—statements that are logically inferred but not explicitly stated. Most current NLI datasets are focused on explicit entailments, making the models insufficiently equipped to deal with scenarios where meaning is indirectly expressed. This limitation bars the development of applications such as conversational AI, summarization, and context-sensitive decision-making, where the ability to infer unspoken implications is crucial. To mitigate this shortcoming, a dataset and approach that systematically incorporates implied entailments in NLI tasks are needed.

Current NLI benchmarks like SNLI, MNLI, ANLI, and WANLI are largely dominated by explicit entailments, with implied entailments making up a negligible proportion of the dataset. Therefore, state-of-the-art models trained on these datasets tend to mislabel implied entailments as neutral or contradictory. Previous efforts in introducing an understanding of implicature have been focused on structured inputs like indirect question-answering or pre-defined logical relations, which do not generalize to free-form reasoning settings. Even large models like GPT-4 exhibit a significant performance gap between explicit and implicit entailment detection, which calls for a more comprehensive approach.

Google Deepmind and University of Pennsylvania researchers have proposed the Implied NLI (INLI) dataset to bridge the gap between the explicit and implicit entailments in natural language inference (NLI) models. Their paper proposes a systematic method of incorporating implied meaning in NLI training using structured implicature frameworks from current datasets such as LUDWIG, CIRCA, NORMBANK, and SOCIALCHEM to transform these frameworks into pairs of ⟨premise, implied entailment⟩. In addition, each premise is also paired with explicit entailments, neutral hypotheses, and contradictions to create an inclusive dataset for model training. A groundbreaking few-shot prompting method using Gemini-Pro ensures the generation of high-quality implicit entailments while, concurrently, reducing annotation expenses and ensuring data integrity. Incorporating implicit meaning in NLI tasks enables the differentiation between explicit and implicit entailments by models with higher precision.

The creation of the INLI dataset is a two-stage procedure. First, existing structured datasets with implicatures such as indirect replies and social norms are restructured into an ⟨implied entailment, premise⟩ format. In stage two, to ensure the strength of the dataset, explicit entailments, neutral statements, and contradictions are generated through controlled manipulation of the implied entailments. The dataset comprises 40,000 hypotheses (implied, explicit, neutral, and contradictory) for 10,000 premises, offering a diverse and balanced training set. Fine-tuning experiments using T5-XXL models employ a range of learning rates (1e-6, 5e-6, 1e-5) over 50,000 training steps to improve the identification of implicit entailments.

Models fine-tuned on INLI show a dramatic improvement in detecting implied entailments, with an optimal accuracy of 92.5% compared to 50–71% accuracy for models fine-tuned on typical NLI datasets. Fine-tuned models generalize well to unseen datasets with high accuracy, scoring 94.5% on NORMBANK and 80.4% on SOCIALCHEM, establishing the robustness of INLI on varied domains. Furthermore, hypothesis-only baselines prove that models fine-tuned on INLI leverage both premise and hypothesis for inference, decreasing the likelihood of shallow pattern learning. These results establish the robustness of INLI in bridging explicit and implicit entailments, and in turn, substantially improving AI’s capacity for refined human communication.

This paper makes significant contributions to NLI by proposing the Implied NLI (INLI) dataset, which systematically introduces implied meaning to inference tasks. Employing structured implicature frames and alternative hypothesis generation, this approach improves model accuracy for detecting implicit entailments and facilitates improved generalization across domains. With strong empirical evidence to establish its robustness, INLI establishes a new benchmark for training AI models to identify implicit meaning, leading to more nuanced and context-aware natural language understanding.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 70k+ ML SubReddit.

Meet IntellAgent: An Open-Source Multi-Agent Framework to Evaluate Complex Conversational AI System (Promoted)

The post Can AI Understand Subtext? A New AI Approach to Natural Language Inference appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

自然语言推理 隐式含义 INLI数据集 深度学习
相关文章