hh.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Continuous transfer of neural network representational similarity for incremental learning
Chinese Academy of Sciences, Beijing, China; University of Chinese Academy of Sciences, Beijing, China.
Chinese Academy of Sciences, Beijing, China; University of Chinese Academy of Sciences, Beijing, China.ORCID iD: 0000-0001-9668-2883
Chinese Academy of Sciences, Beijing, China; University of Chinese Academy of Sciences, Beijing, China; Zhongke Ruitu Technology Co., Ltd, Beijing, China.ORCID iD: 0000-0001-7897-1673
Chinese Academy of Sciences, Beijing, China.
Show others and affiliations
2023 (English)In: Neurocomputing, ISSN 0925-2312, E-ISSN 1872-8286, Vol. 545, article id 126300Article in journal (Refereed) Published
Abstract [en]

The incremental learning paradigm in machine learning has consistently been a focus of academic research. It is similar to the way in which biological systems learn, and reduces energy consumption by avoiding excessive retraining. Existing studies utilize the powerful feature extraction capabilities of pre-trained models to address incremental learning, but there remains a problem of insufficient utilization of neural network feature knowledge. To address this issue, this paper proposes a novel method called Pre-trained Model Knowledge Distillation (PMKD) which combines knowledge distillation of neural network representations and replay. This paper designs a loss function based on centered kernel alignment to transfer neural network representations knowledge from the pre-trained model to the incremental model layer-by-layer. Additionally, the use of memory buffer for Dark Experience Replay helps the model retain past knowledge better. Experiments show that PMKD achieved superior performance on various datasets and different buffer sizes. Compared to other methods, our class incremental learning accuracy reached the best performance. The open-source code is published athttps://github.com/TianSongS/PMKD-IL. © 2023 The Author(s)

Place, publisher, year, edition, pages
Amsterdam: Elsevier, 2023. Vol. 545, article id 126300
Keywords [en]
Incremental learning, Knowledge distillation, Neural network representation, Pre-trained model
National Category
Natural Language Processing
Identifiers
URN: urn:nbn:se:hh:diva-51402DOI: 10.1016/j.neucom.2023.126300ISI: 001001824300001Scopus ID: 2-s2.0-85159353460OAI: oai:DiVA.org:hh-51402DiVA, id: diva2:1788039
Note

Funding: National Natural Science Foundation of China (No. 61901436)

Available from: 2023-08-15 Created: 2023-08-15 Last updated: 2025-02-07Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Tiwari, Prayag

Search in DiVA

By author/editor
Li, WeijunNing, XinTiwari, Prayag
By organisation
School of Information Technology
In the same journal
Neurocomputing
Natural Language Processing

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 37 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf