首页 AI学术青年与开发者社区

增量学习-论文速读

艾蜜莉  

2018-05-07

阅读时间8:51-9:39

 

iCaRL: Incremental Classifier and Representation Learning


Abstract: A major open problem on the road to artificial intelligence is the development of incrementally learning systems that learn about more and more concepts over time from a stream of data. In this work, we introduce a new training strategy, iCaRL, that allows learning in such a class-incremental way: only the training data for a small number of classes has to be present at the same time and new classes can be added progressively. iCaRL learns strong classifiers and a data representation simultaneously. This distinguishes it from earlier works that were fundamentally limited to fixed data representations and therefore incompatible with deep learning architectures. We show by experiments on CIFAR-100 and ImageNet ILSVRC 2012 data that iCaRL can learn many classes incrementally over a long period of time where other strategies quickly fail.

代码实现:https://github.com/Khurramjaved96/incremental-learning

平台:Pytorch

以增量的方式学习,并且可以逐渐增加新类,可以在很长时间逐步学习新类。

 

From N to N+1: Multiclass Transfer Incremental Learning

Since the seminal work of Thrun [16], the learning to learn paradigm has been defined as the ability of an agent to improve its performance at each task with experience, with the number of tasks. Within the object categorization domain, the visual learning community has actively declined this paradigm in the transfer learning setting. Almost all proposed methods focus on category detection problems, addressing how to learn a new target class from few samples by leveraging over the known source. But if one thinks of learning over multiple tasks, there is a need for multiclass transfer learning algorithms able to exploit previous source knowledge when learning a new class, while at the same time optimizing their overall performance. This is an open challenge for existing transfer learning algorithms. The contribution of this paper is a discriminative method that addresses this issue, based on a Least-Squares Support Vector Machine formulation. Our approach is designed to balance between transferring to the new class and preserving what has already been learned on the source models. Extensive experiments on subsets of publicly available datasets prove the effectiveness of our approach.

基于最小二乘支持向量机解决在学习新类时利用以前的源知识,同时优化其整体性能的问题

 

Recent Advances in Zero-shot Recognition


Abstract: With the recent renaissance of deep convolution neural networks, encouraging breakthroughs have been achieved on the supervised recognition tasks, where each class has sufficient training data and fully annotated training data. However, to scale the recognition to a large number of classes with few or now training samples for each class remains an unsolved problem. One approach to scaling up the recognition is to develop models capable of recognizing unseen categories without any training instances, or zero-shot recognition/ learning. This article provides a comprehensive review of existing zero-shot recognition techniques covering various aspects ranging from representations of models, and from datasets and evaluation settings. We also overview related recognition tasks including one-shot and open set recognition which can be used as natural extensions of zero-shot recognition when limited number of class samples become available or when zero-shot recognition is implemented in a real-world setting. Importantly, we highlight the limitations of existing approaches and point out future research directions in this existing new research area.

零点识别技术,在没有任何训练实例的情况下识别未发现的类别

 

Incremental Learning of Object Detectors without Catastrophic Forgetting

Despite their success for object detection, convolutional neural networks are ill-equipped for incremental learning, i.e., adapting the original model trained on a set of classes to additionally detect objects of new classes, in the absence of the initial training data. They suffer from “catastrophic forgetting”–an abrupt degradation of performance on the original set of classes, when the training objective is adapted to the new classes. We present a method to address this issue, and learn object detectors incrementally, when neither the original training data nor annotations for the original classes in the new training set are available. The core of our proposed solution is a loss function to balance the interplay between predictions on the new classes and a new distillation loss which minimizes the discrepancy between responses for old classes from the original and the updated networks. This incremental learning can be performed multiple times, for a new set of classes in each step, with a moderate drop in performance compared to the baseline network trained on the ensemble of data. We present object detection results on the PASCAL VOC 2007 and COCO datasets, along with a detailed empirical analysis of the approach.

在新类别上进行训练逐步调整原始模型以检测新类别时,表现出灾难性的遗忘,在原始类别上的性能突然退化。本文提出的平衡方法使得增量学习可以多次执行,与在数据集合上训练的基线网络相比,性能会有所下降。

Learning multiple visual domains with residual adapters

Abstract: There is a growing interest in learning data representations that work well for many different types of problems and data. In this paper, we look in particular at the task of learning a single visual representation that can be successfully utilized in the analysis of very different types of images, from dog breeds to stop signs and digits. Inspired by recent work on learning networks that predict the parameters of another, we develop a tunable deep network architecture that, by means of adapter residual modules, can be steered on the fly to diverse visual domains. Our method achieves a high degree of parameter sharing while maintaining or even improving the accuracy of domain-specific representations. We also introduce the Visual Decathlon Challenge, a benchmark that evaluates the ability of representations to capture simultaneously ten very different visual domains and measures their ability to recognize well uniformly.

学习对多种问题和数据都能表现很好的数据表示引发了越来越多的兴趣。本文可以分析不同类型的图像,从狗品种到停止标志和数字,开发了一种可调节的深度网络架构,通过适配器残留模块,可以实时引导到不同的视觉领域。还介绍了视觉十项全能挑战赛,这是一个基准,用于评估表征能够同时捕捉十个非常不同的视觉领域,并衡量他们的统一识别能力。

 

总结:在物体识别过程中逐步新增类别,深度学习方法目前还不实用,传统实现方式更靠谱?

0条评论