MarkTechPost@AI 2024年10月01日
RanDumb: A Simple Yet Powerful AI Approach to Exemplar-Free Continual Learning
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

RanDumb是一种新颖的无示例连续学习方法,它利用随机傅里叶特征和线性分类器来创建有效的表示,无需存储示例或进行频繁更新。该方法通过将原始输入像素投影到高维特征空间,使用随机傅里叶变换来近似径向基函数(RBF)内核,从而实现有效的表示学习。这种固定随机投影之后是一个简单的线性分类器,根据变换后的特征的最近类均值对它们进行分类。RanDumb在多个连续学习基准测试中表现出色,在MNIST、CIFAR-10和CIFAR-100等数据集上取得了优于现有方法的性能,证明了其在处理连续在线和离线学习场景中的鲁棒性,无需存储示例或采用复杂的训练策略。

🎉 RanDumb 是一种基于随机特征的无示例连续学习方法,它利用随机傅里叶变换和线性分类器来创建有效的表示,无需存储示例或进行频繁更新。

💡 RanDumb 通过将原始输入像素投影到高维特征空间,使用随机傅里叶变换来近似径向基函数(RBF)内核,从而实现有效的表示学习。这种固定随机投影之后是一个简单的线性分类器,根据变换后的特征的最近类均值对它们进行分类。

📊 RanDumb 在多个连续学习基准测试中表现出色,在 MNIST、CIFAR-10 和 CIFAR-100 等数据集上取得了优于现有方法的性能,证明了其在处理连续在线和离线学习场景中的鲁棒性,无需存储示例或采用复杂的训练策略。

🚀 RanDumb 的性能扩展到包含预训练特征提取器的场景。在 TinyImageNet 等复杂数据集上,该方法使用随机投影之上的简单线性分类器,实现了接近最先进的性能。该方法成功地将性能差距缩小到联合分类器,显著优于大多数连续微调和提示微调策略。

💪 RanDumb 在低示例场景中表现出显著的性能提升,在这些场景中,数据存储受限或不可用。例如,在 CIFAR-100 数据集上的离线连续学习中,RanDumb 优于之前领先的方法 4%。

Continual learning is a rapidly evolving area of research that focuses on developing models capable of learning from sequentially arriving data streams, similar to human learning. It addresses the challenges of adapting to new information while retaining previously acquired knowledge. This field is particularly relevant in scenarios where models must perform well on multiple tasks over extended periods, such as real-world applications with non-stationary data and limited computational resources. Unlike traditional machine learning, where models are trained on static datasets, continual learning requires models to adapt dynamically to new data while managing memory and computational efficiency.

A significant issue in continual learning is the problem of “catastrophic forgetting,” where neural networks lose the ability to recall previously learned tasks when exposed to new ones. This phenomenon is especially problematic when models need help to store or revisit old data, making it difficult to balance learning stability and model adaptability. The inability to effectively integrate new information without sacrificing the performance of prior knowledge remains a major hurdle. Researchers have been trying to design solutions that address this limitation. Yet, many existing methods fail to achieve the desired results in exemplar-free scenarios where no previous data samples can be stored for future reference.

Existing methods to tackle catastrophic forgetting generally involve joint training of representations alongside classifiers or using experience replay and regularization techniques. These approaches, however, assume that representations derived from continually learned neural networks will naturally outperform predefined random functions, as observed in standard deep learning setups. The core issue is that these methods are not evaluated under the constraints of continual learning. For instance, models often cannot be updated sufficiently in online continual learning scenarios before data is discarded. This results in suboptimal representations and reduced classification accuracy when dealing with new data streams.

Researchers from the University of Oxford, IIIT Hyderabad, and Apple have developed a novel approach called RanDumb. The method utilizes a combination of random Fourier features and a linear classifier to create effective representations for classification without the need for storing exemplars or performing frequent updates. RanDumb’s mechanism is straightforward—it projects raw input pixels into a high-dimensional feature space using a random Fourier transform, which approximates the Radial Basis Function (RBF) Kernel. This fixed random projection is followed by a simple linear classifier that classifies the transformed features based on their nearest class means. This method outperforms many existing techniques by eliminating the need for fine-tuning or complex neural network updates, making it highly suitable for exemplar-free continual learning.

RanDumb operates by embedding the input data into a high-dimensional space, decorrelating the features using Mahalanobis distance and cosine similarity for accurate classification. Unlike traditional methods that update representations alongside classifiers, RanDumb uses a fixed random transform for embedding. It only requires online updates to the covariance matrix and class means, allowing it to handle new data as it arrives efficiently. The approach also bypasses the need for memory buffers, making it an ideal solution for low-resource environments. Furthermore, the method retains computational simplicity by operating on one sample at a time, ensuring scalability even with large datasets.

Experimental evaluations demonstrate that RanDumb consistently performs well across multiple continual learning benchmarks. For example, on the MNIST dataset, RanDumb achieved an accuracy of 98.3%, surpassing existing methods by 5-15% margins. In CIFAR-10 and CIFAR-100 benchmarks, RanDumb recorded accuracies of 55.6% and 28.6%, respectively, outperforming state-of-the-art methods that rely on storing previous samples. The results highlight the method’s robustness in handling continual online and offline learning scenarios without storing exemplars or employing complex training strategies. In particular, RanDumb matched or exceeded the performance of joint training on many benchmarks, bridging 70-90% of the performance gap between constrained continual learning and unconstrained joint learning.

Moreover, RanDumb’s efficiency extends to scenarios that incorporate pretrained feature extractors. When applied to complex datasets like TinyImageNet, the proposed method achieved near state-of-the-art performance using a simple linear classifier on top of random projections. The approach managed to bridge the performance gap to joint classifiers by up to 90%, significantly outperforming most continual fine-tuning and prompt-tuning strategies. Further, the method shows a marked performance gain in low-exemplar scenarios where data storage is restricted or unavailable. For example, RanDumb outperformed previous leading methods by 4% on the CIFAR-100 dataset in offline continual learning.

In conclusion, the RanDumb approach redefines the assumptions surrounding effective representation learning in continual learning. Its random feature-based methodology proves to be a simpler yet more powerful solution for representation learning, challenging the conventional reliance on complex neural network updates. The research addresses the limitations of current continual learning methods and opens up new avenues for developing efficient and scalable solutions in exemplar-free and resource-constrained environments. By leveraging the power of random embeddings, RanDumb paves the way for future advancements in continual learning, especially in online learning scenarios where data and computational resources are limited.


Check out the Paper and GitHub. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter..

Don’t Forget to join our 52k+ ML SubReddit.

We are inviting startups, companies, and research institutions who are working on small language models to participate in this upcoming ‘Small Language Models’ Magazine/Report by Marketchpost.com. This Magazine/Report will be released in late October/early November 2024. Click here to set up a call!

The post RanDumb: A Simple Yet Powerful AI Approach to Exemplar-Free Continual Learning appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

连续学习 无示例学习 随机傅里叶特征 线性分类器 RanDumb
相关文章