MarkTechPost@AI 2024年07月08日
Google Researchers Propose a Formal Boosting Machine Learning Algorithm for Any Loss Function Whose Set of Discontinuities has Zero Lebesgue Measure
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

谷歌研究团队提出了一种新的机器学习算法SECBOOST,该算法能够处理具有零勒贝格测度不连续点的任意损失函数。这一技术不仅解决了局部最小值问题,还能处理在其区域内部分稳定值的损失函数,为未来的提升研究和应用提供了新的可能性。

🔍 SECBOOST技术通过引入量子微积分的策略,避免了在提升过程中使用导数,从而能够处理非凸、非可微、非Lipschitz或非连续的损失函数。这一方法与传统的零阶优化解决方案有显著不同,后者在证明收敛性时通常假设损失函数具有更多的性质,如凸性、可微性等。

🚀 SECBOOST技术在更广泛的上下文中应用时,揭示了两个额外的领域,通过有意识的设计决策可以在更多的轮次中保持假设。这不仅解决了局部最小值的问题,还能处理在其区域内部分稳定值的损失函数,显示了该技术的巨大潜力。

📈 研究表明,提升技术在某些方面优于最新的零阶优化进展。这是因为为了实现提升兼容的收敛,损失函数仅被假设满足一些在类似分析中常用的典型假设。尽管这一问题在此情况下需要解决,例如为了有效地优化偏移预言机,但最近的零阶优化进展也已经实现了一些重要的设计技巧来实现此类算法。

As a very effective machine learning ML-born optimization setting, boosting requires one to efficiently learn arbitrarily good models using a weak learner oracle, which provides classifiers that perform marginally better than random guessing. Although the original boosting model did not necessitate first-order loss information, the decades-long history of boosting has rapidly transformed it into a first-order optimization setting, with some even incorrectly defining it as such. This is a significant difference with gradient-based optimization.

The term “zeroth order optimization” can describe a group of optimization methods that skip over using gradient information to determine a function’s minimum and maximum values. These techniques shine in cases where the function is either noisy or non-differentiable or where computing the gradient would be prohibitively expensive or impractical. In contrast, the search for the best solution in zeroth order optimization is guided entirely by function evaluations.

There have been few investigations into boosting, even though ML has witnessed a significant uptick in zeroth order optimization across numerous settings and algorithms in recent years. The question is highly pertinent, as boosting has rapidly developed into a method that necessitates first-order knowledge of the optimal loss. Boosting lowered to this first-order setting is also rather typical. A weak learner that could provide classifiers that were distinct from random guessing was originally required by the boosting model rather than first-order loss information. With zeroth-order optimization becoming more popular in machine learning ML, it’s important to know if differentiability is necessary for boosting, which loss functions can be boosted with a weak learner, and how boosting compares to the recent formal progress on bringing gradient descent to zeroth-order optimization.

Google’s research team aims to provide a formal boosting technique to handle loss functions with sets of discontinuities with zero Lebesgue measure. Any stored loss function would, in reality, satisfy this criterion with conventional floating-point encoding. Theoretically, the researchers include losses that are not necessarily convex, differentiable, Lipschitz, or continuous. Classical zeroth-order optimization solutions differ significantly in this regard; while their algorithms are zeroth-order, the assumptions made about the loss in their proof of convergence—including convexity, differentiability (once or twice), Lipschitzness, and so on—are far more extensive. They employ or expand upon strategies from quantum calculusℎ, some of which seem to be commonplace in zeroth-order optimization research, to sidestep the usage of derivatives in boosting.

The proposed SECBOOST technique, when applied to a broader context, uncovers two additional areas where deliberate design decisions can be leveraged to maintain assumptions throughout a stronger number of rounds. This not only addresses the issue of local minima but also manages losses that exhibit stable values over portions of their area. The potential of the SECBOOST technique is significant, offering hope for the future of boosting research and application.

Based on the findings, boosting is better than the latest advancements in zeroth-order optimization. This is because, to achieve boosting-compliant convergence, the loss was only assumed to satisfy some of the typical assumptions used in such analyses. While this issue requires fixing in this situation—for example, to optimize the offset oracle efficiently—recent developments in zeroth-order optimization have also accomplished significant design tricks for implementing such algorithms. The team hasn’t resolved this issue yet. Still, in the appendix, the community can find some mock experiments that a simple implementation can accomplish, suggesting that SecBoost can optimize “exotic” types of losses. 


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter

Join our Telegram Channel and LinkedIn Group.

If you like our work, you will love our newsletter..

Don’t Forget to join our 46k+ ML SubReddit

The post Google Researchers Propose a Formal Boosting Machine Learning Algorithm for Any Loss Function Whose Set of Discontinuities has Zero Lebesgue Measure appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

机器学习 提升算法 零阶优化 SECBOOST 谷歌研究
相关文章