hh.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Few-Shot Class-Incremental Learning for Medical Time Series Classification
Nanjing University of Information Science And Technology, Nanjing, China.ORCID iD: 0000-0002-4221-0327
Nanjing University of Information Science And Technology, Nanjing, China.
The Chinese University of Hong Kong, Shenzhen, China.ORCID iD: 0000-0002-1501-9914
Halmstad University, School of Information Technology.ORCID iD: 0000-0002-2851-4260
2024 (English)In: IEEE journal of biomedical and health informatics, ISSN 2168-2194, E-ISSN 2168-2208, Vol. 28, no 4, p. 1872-1882Article in journal (Refereed) Published
Abstract [en]

Continuously analyzing medical time series as new classes emerge is meaningful for health monitoring and medical decision-making. Few-shot class-incremental learning (FSCIL) explores the classification of few-shot new classes without forgetting old classes. However, little of the existing research on FSCIL focuses on medical time series classification, which is more challenging to learn due to its large intra-class variability. In this paper, we propose a framework, the Meta self-Attention Prototype Incrementer (MAPIC) to address these problems. MAPIC contains three main modules: an embedding encoder for feature extraction, a prototype enhancement module for increasing inter-class variation, and a distance-based classifier for reducing intra-class variation. To mitigate catastrophic forgetting, MAPIC adopts a parameter protection strategy in which the parameters of the embedding encoder module are frozen at incremental stages after being trained in the base stage. The prototype enhancement module is proposed to enhance the expressiveness of prototypes by perceiving inter-class relations using a self-attention mechanism. We design a composite loss function containing the sample classification loss, the prototype non-overlapping loss, and the knowledge distillation loss, which work together to reduce intra-class variations and resist catastrophic forgetting. Experimental results on three different time series datasets show that MAPIC significantly outperforms state-of-the-art approaches by 27.99%, 18.4%, and 3.95%, respectively. IEEE

Place, publisher, year, edition, pages
Piscataway, N.J.: IEEE, 2024. Vol. 28, no 4, p. 1872-1882
Keywords [en]
Brain modeling, Few-shot class-incremental learning, Health monitoring, Medical decision-making, Medical diagnostic imaging, Medical time series classification, Power capacitors, Prototypes, Task analysis, Time series analysis, Training
National Category
Natural Language Processing
Identifiers
URN: urn:nbn:se:hh:diva-51964DOI: 10.1109/JBHI.2023.3247861ISI: 001197865400008Scopus ID: 2-s2.0-85149384773OAI: oai:DiVA.org:hh-51964DiVA, id: diva2:1811496
Available from: 2023-11-13 Created: 2023-11-13 Last updated: 2025-02-07Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Tiwari, Prayag

Search in DiVA

By author/editor
Sun, LeWang, BenyouTiwari, Prayag
By organisation
School of Information Technology
In the same journal
IEEE journal of biomedical and health informatics
Natural Language Processing

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 31 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf