MarkTechPost@AI 2024年10月20日
Embed-then-Regress: A Versatile Machine Learning Approach for Bayesian Optimization Using String-Based In-Context Regression
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

传统的贝叶斯优化方法通常依赖于回归模型来预测固定搜索空间内解决方案的性能,但许多回归方法由于建模假设和输入限制而具有任务特异性。最近,大型语言模型(LLM)的进展为克服这些限制提供了希望,通过将搜索空间候选嵌入为字符串,使更灵活、通用的回归器能够跨任务进行泛化,并超越传统回归方法的限制。这项研究提出了一种名为“Embed-then-Regress”的范式,它利用来自预训练语言模型的字符串嵌入进行上下文回归,将所有输入转换为字符串表示,从而实现针对各种任务的通用回归,包括合成、组合和超参数优化。

😁 **Embed-then-Regress方法**:通过将搜索空间候选嵌入为字符串,克服了传统回归方法的限制,使更灵活、通用的回归器能够跨任务进行泛化。

😊 **字符串嵌入**:该方法将所有输入转换为字符串表示,并使用LLM将字符串映射到固定长度的向量,以便使用基于张量的回归器(如Transformer模型)。

😉 **预训练和不确定性感知预测**:在大型离线数据集上进行预训练,使模型能够对看不见的目标进行不确定性感知的预测。

😄 **实验结果**:实验结果表明,该方法在各种任务(包括合成、组合和超参数优化任务)中有效,并且能够处理优化场景中的连续和分类参数的混合。

🥳 **未来方向**:未来的研究可以集中在开发一个通用的上下文回归模型上,方法是在各种领域进行预训练,并改进体系结构方面,例如学习Transformer输出的聚合方法。

Bayesian Optimization, widely used in experimental design and black-box optimization, traditionally relies on regression models for predicting the performance of solutions within fixed search spaces. However, many regression methods are task-specific due to modeling assumptions and input constraints. This issue is especially prevalent in learning-based regression, which depends on fixed-length tensor inputs. Recent advancements in LLMs show promise in overcoming these limitations by embedding search space candidates as strings, enabling more flexible, universal regressors to generalize across tasks and expand beyond the constraints of traditional regression methods.

Bayesian Optimization uses regressors to solve black-box optimization problems by balancing exploration and exploitation. Traditionally dominated by Gaussian Process (GP) regressors, recent efforts have focused on improving GP hyperparameters through pretraining or feature engineering. While neural network approaches like Transformers offer more flexibility, they are limited by fixed input dimensions, restricting their application to tasks with structured inputs. Recent advances propose embedding string representations of search space candidates for greater task flexibility. This approach enables efficient, trainable regressors to handle diverse inputs, longer sequences, and precise predictions across varying scales, improving optimization performance.

Researchers from UCLA, Google DeepMind, and Google propose the “Embed-then-Regress” paradigm for in-context regression using string embeddings from pretrained language models. Converting all inputs into string representations enables general-purpose regression for Bayesian Optimization across diverse tasks like synthetic, combinatorial, and hyperparameter optimization. Their framework uses LLM-based embeddings to map strings to fixed-length vectors for tensor-based regressors, such as Transformer models. Pretraining on large offline data sets allows uncertainty-aware predictions for unseen objectives. The framework, enhanced with explore-exploit techniques, delivers results comparable to state-of-the-art Gaussian Process-based optimization algorithms.

The method uses an embedding-based regressor for Bayesian optimization, mapping string inputs to fixed-length vectors via a language model. These embeddings are processed by a Transformer to predict outcomes, forming an acquisition function to balance exploration and exploitation. The model, pretrained on offline tasks, uses historical data to make uncertainty-aware predictions. During inference, a mean and deviation output guides optimization. The approach is computationally efficient, using a T5-XL encoder and a smaller Transformer, requiring moderate GPU resources. This framework achieves scalable predictions while maintaining a low inference cost through efficient Transformers and embeddings.

The experiment demonstrates the versatility of the Embed-then-Regress method across a wide range of tasks, focusing on its broad applicability rather than optimizing for specific domains. The algorithm was evaluated on various problems, including synthetic, combinatorial, and hyperparameter optimization tasks, with performance averaged over multiple runs. The results show that the method effectively handles a mix of continuous and categorical parameters in optimization scenarios. The approach highlights its potential in diverse optimization settings, offering a flexible solution for different problem types without needing domain-specific adjustments.

In conclusion, the Embed-then-Regress method showcases the flexibility of string-based in-context regression for Bayesian Optimization across diverse problems, achieving results comparable to standard GP methods while handling complex data types like permutations and combinations. Future research could focus on developing a universal in-context regression model by pretraining across various domains and improving architectural aspects, such as learning aggregation methods for Transformer outputs. Additional applications could include optimizing prompts and code search, which rely on less efficient algorithms. Exploring the use of this approach in process-based reward modeling and stateful environments in language modeling is also promising.


Check out the Paper and GitHub. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter.. Don’t Forget to join our 50k+ ML SubReddit.

[Upcoming Live Webinar- Oct 29, 2024] Learn how to increase inference throughput by 4x and reduce serving costs by 50% with Turbo LoRA, FP8 and GPU Autoscaling (Promoted)

The post Embed-then-Regress: A Versatile Machine Learning Approach for Bayesian Optimization Using String-Based In-Context Regression appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

贝叶斯优化 机器学习 大型语言模型 字符串嵌入 上下文回归
相关文章