MarkTechPost@AI 04月14日
Foundation Models No Longer Need Prompts or Labels: EPFL Researchers Introduce a Joint Inference Framework for Fully Unsupervised Adaptation Using Fine-Tuning and In-Context Learning
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

EPFL的研究人员开发了一种联合推理框架,使Foundation模型能够进行完全无监督的自适应。该框架通过无监督微调和无监督上下文学习,使得这些大型模型无需手动提示或标注数据即可提高性能。研究结果表明,在数学推理和自然语言处理等任务中,该方法实现了显著的性能提升,甚至在某些任务上超越了有监督微调的效果。这项研究为Foundation模型在实际应用中提供了更广泛的可能性,减少了对人工干预的依赖。

💡 基础模型通常是经过大量文本和图像数据训练的大型神经网络,它们在处理语言和视觉任务方面具有显著优势。这些模型擅长通用任务,通过预训练知识实现跨领域的泛化。

⚙️ 传统的Foundation模型在适应新任务时,通常需要手工制作提示或标注示例,这增加了工作量。手动制作提示需要反复试验,而收集标注数据则耗时且昂贵,限制了模型在零样本设置下的可用性。

✨ EPFL的研究人员提出了一个联合推理框架,支持无监督自适应。该框架通过无监督微调和无监督上下文学习,使模型无需真实数据或手动提示即可进行协调预测,从而提高了模型的准确性。

📈 无监督微调通过让模型利用自身的反馈迭代地改进预测。它使用LoRA进行高效的权重更新,并引入正则化步骤以避免出现所有输入都预测相同答案的简单解。无监督上下文学习则通过使用先前生成的输出来模拟标签,在无需人工标注的情况下,通过多次迭代来优化预测。

🚀 研究结果表明,无监督方法带来了显著的性能提升。例如,在GSM8K数学推理数据集上,无监督上下文学习将Qwen2.5-Math模型的性能提高了39.2%。在13个自然语言处理任务中,无监督微调使Llama-3.1-8B模型的平均准确度提高了23%,并在其中6个任务中达到了与完全监督微调相同的水平。

Foundation models, often massive neural networks trained on extensive text and image data, have significantly shifted how artificial intelligence systems handle language and vision tasks. These models are not designed for a single task but generalize across a wide variety by leveraging their pretraining knowledge. Once trained, they can generate coherent responses, classify images, or solve problems without needing new task-specific training. Their scalability and reuse across domains make them a cornerstone of AI development.

Despite their broad capabilities, a persistent issue lies in how these models are adapted for new, unseen tasks. In most scenarios, achieving strong performance requires providing them with handcrafted prompts or labeled examples that guide the model on how to behave. This process, however, introduces overhead, as crafting prompts involves trial and error, and collecting labeled examples can be expensive and time-consuming. Moreover, in real-world applications, such support data may not always be readily available, limiting the usability of foundation models in zero-shot settings.

Several strategies have been used to bridge this gap between generality and task-specific performance. In-context learning enables models to mimic a task by including example input-output pairs during inference, while supervised fine-tuning adjusts model weights using labeled data. Another method, prompt engineering, involves crafting prompts that steer the model toward desired outputs. Though these tools have been successful in boosting performance, each relies on external support—either human input or labeled data—making them less viable in completely unsupervised settings.

Swiss Federal Institute of Technology Lausanne (EPFL) researchers introduced a joint inference framework that supports unsupervised adaptation. This framework enables foundation models to perform coordinated predictions over multiple inputs without requiring ground truth data or manual prompts. The research team presented two specific techniques under this framework: unsupervised fine-tuning and unsupervised in-context learning. These methods allow models, including closed-weight ones like GPT-4, to improve accuracy without external guidance.

The approach of unsupervised fine-tuning works by letting the model iteratively improve its predictions using only its feedback. It formulates an optimization objective where predictions for a batch of inputs are generated together, and their joint probability is maximized. This method uses LoRA (Low-Rank Adaptation) for efficient weight updates and introduces a regularization step to avoid trivial solutions, such as predicting the same answer for all inputs. The researchers developed unsupervised in-context learning for situations where weight access isn’t available, such as with GPT-4. This method mimics the effect of labeled ICL by using previously generated outputs as pseudo-labels, refining predictions over multiple iterations without human annotations. Each iteration involves conditioning the model on prior examples and developing a more accurate answer, simulating a supervised learning loop through self-generated data.

The performance improvements from these unsupervised methods were substantial. On the GSM8K dataset, designed for math reasoning, unsupervised ICL applied to the Qwen2.5-Math model achieved a 39.2% absolute improvement over the standard zero-shot baseline. Similarly, for the Llama-3.1-8B model tested across 13 natural language processing tasks, unsupervised fine-tuning delivered a 23% average gain in accuracy. It matched the performance of fully supervised fine-tuning in 6 out of the 13 tasks. In vision-language tasks, unsupervised ICL also demonstrated strong results—showing a 23% gain on the Food101 dataset and significant improvements across other benchmarks. The research even extended to GPT-4o, a closed-weight model, where a 3% improvement was observed on ImageNet, reinforcing the framework’s versatility.

This work reveals a meaningful shift in how foundation models can adapt. The researchers successfully addressed the core limitation—reliance on labeled data and manual configuration—by introducing a robust and scalable self-supervised strategy. Their joint inference framework is a practical, generalizable approach that redefines the boundaries of unsupervised learning for large-scale AI models.


Check out Paper and Project. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 85k+ ML SubReddit.

The post Foundation Models No Longer Need Prompts or Labels: EPFL Researchers Introduce a Joint Inference Framework for Fully Unsupervised Adaptation Using Fine-Tuning and In-Context Learning appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Foundation模型 无监督学习 自适应 EPFL
相关文章