A survey on few-shot class-incremental learningShow others and affiliations
2024 (English)In: Neural Networks, ISSN 0893-6080, E-ISSN 1879-2782, Vol. 169, p. 307-324Article, review/survey (Refereed) Published
Abstract [en]
Large deep learning models are impressive, but they struggle when real-time data is not available. Few-shot class-incremental learning (FSCIL) poses a significant challenge for deep neural networks to learn new tasks from just a few labeled samples without forgetting the previously learned ones. This setup can easily leads to catastrophic forgetting and overfitting problems, severely affecting model performance. Studying FSCIL helps overcome deep learning model limitations on data volume and acquisition time, while improving practicality and adaptability of machine learning models. This paper provides a comprehensive survey on FSCIL. Unlike previous surveys, we aim to synthesize few-shot learning and incremental learning, focusing on introducing FSCIL from two perspectives, while reviewing over 30 theoretical research studies and more than 20 applied research studies. From the theoretical perspective, we provide a novel categorization approach that divides the field into five subcategories, including traditional machine learning methods, meta learning-based methods, feature and feature space-based methods, replay-based methods, and dynamic network structure-based methods. We also evaluate the performance of recent theoretical research on benchmark datasets of FSCIL. From the application perspective, FSCIL has achieved impressive achievements in various fields of computer vision such as image classification, object detection, and image segmentation, as well as in natural language processing and graph. We summarize the important applications. Finally, we point out potential future research directions, including applications, problem setups, and theory development. Overall, this paper offers a comprehensive analysis of the latest advances in FSCIL from a methodological, performance, and application perspective. © 2023 The Author(s)
Place, publisher, year, edition, pages
Oxford: Elsevier, 2024. Vol. 169, p. 307-324
Keywords [en]
Catastrophic forgetting, Class-incremental learning, Few-shot learning, Overfitting, Performance evaluation
National Category
Computer Vision and Robotics (Autonomous Systems)
Identifiers
URN: urn:nbn:se:hh:diva-52064DOI: 10.1016/j.neunet.2023.10.039Scopus ID: 2-s2.0-85175341096OAI: oai:DiVA.org:hh-52064DiVA, id: diva2:1812997
Note
Funding: This work is supported by the National Natural Science Foundation of China (Grant No. 62373343) and the Beijing Natural Science Foundation, China (No. L233036).
2023-11-172023-11-172023-11-17Bibliographically approved