hh.sePublications
12 2 of 2
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Advancing Meta-Learning for Enhanced Generalization Across Diverse Tasks
Halmstad University, School of Information Technology.ORCID iD: 0000-0003-0185-5038
2025 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Meta-learning, or learning to learn, is a rapidly evolving area in machine learning that aims to enhance the adaptability and efficiency of learning algorithms. Inspired by the human ability to learn new concepts from limited examples and quickly adapt to unforeseen situations, meta-learning leverages prior experience to prepare models for fast adaptation to new tasks. Unlike traditional machine learning systems, where models are trained for specific tasks, meta-learning frameworks enable models to acquire generalized knowledge during training and efficiently learn new tasks during inference. This ability to generalize from past experiences to new tasks makes meta-learning a key focus in advancing artificial intelligence, offering the potential to create more flexible and efficient AI systems capable of performing well with minimal data.

In this thesis, we begin by formally defining the meta-learning framework, establishing clear terminology, and synthesizing existing work in a comprehensive survey paper. Building on this foundation, we demonstrate how meta-learning can be integrated into various fields to enhance model performance and extend capabilities to few-shot learning scenarios. We show how meta-learning can significantly improve the accuracy and efficiency of transferring knowledge across domains in domain adaptation. In scenarios involving a multimodal distribution of tasks, we develop methods that efficiently learn from and adapt to a wide variety of tasks drawn from different modes within the distribution, ensuring effective adaptation across diverse domains. Our work on personalized federated learning highlights meta-learning's potential to tailor federated learning processes to individual user needs while maintaining privacy and data security. Additionally, we address the challenges of continual learning by developing models that continuously integrate new information without forgetting previously acquired knowledge. For time series data analysis, we present meta-learning strategies that automatically learn optimal augmentation techniques, enhancing model predictions and offering robust solutions for real-world applications. Lastly, our pioneering research on unsupervised meta-learning via in-context learning explores innovative approaches for constructing tasks and learning effectively from unlabeled data.

Overall, the contributions of this thesis emphasize the potential of meta-learning techniques to improve performance across diverse research areas and demonstrate how advancements in one area can benefit the field as a whole.

Place, publisher, year, edition, pages
Halmstad: Halmstad University Press, 2025. , p. 46
Series
Halmstad University Dissertations ; 127
Keywords [en]
Meta-learning, Few-shot learning, Domain adaptation, Federated learning, Continual learning, Unsupervised learning, In-context learning
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:hh:diva-55147ISBN: 978-91-89587-71-7 (print)ISBN: 978-91-89587-70-0 (electronic)OAI: oai:DiVA.org:hh-55147DiVA, id: diva2:1924707
Public defence
2025-02-03, S1022, Kristian IV:s väg 3, 30118, Halmstad, Halmstad, 13:00 (English)
Opponent
Supervisors
Available from: 2025-01-08 Created: 2025-01-07 Last updated: 2025-01-08Bibliographically approved
List of papers
1. Advances and Challenges in Meta-Learning: A Technical Review
Open this publication in new window or tab >>Advances and Challenges in Meta-Learning: A Technical Review
Show others...
2024 (English)In: IEEE Transactions on Pattern Analysis and Machine Intelligence, ISSN 0162-8828, E-ISSN 1939-3539, Vol. 46, no 7, p. 4763-4779Article, review/survey (Refereed) Published
Abstract [en]

Meta-learning empowers learning systems with the ability to acquire knowledge from multiple tasks, enabling faster adaptation and generalization to new tasks. This review provides a comprehensive technical overview of meta-learning, emphasizing its importance in real-world applications where data may be scarce or expensive to obtain. The paper covers the state-of-the-art meta-learning approaches and explores the relationship between meta-learning and multi-task learning, transfer learning, domain adaptation and generalization, selfsupervised learning, personalized federated learning, and continual learning. By highlighting the synergies between these topics and the field of meta-learning, the paper demonstrates how advancements in one area can benefit the field as a whole, while avoiding unnecessary duplication of efforts. Additionally, the paper delves into advanced meta-learning topics such as learning from complex multi-modal task distributions, unsupervised metalearning, learning to efficiently adapt to data distribution shifts, and continual meta-learning. Lastly, the paper highlights open problems and challenges for future research in the field. By synthesizing the latest research developments, this paper provides a thorough understanding of meta-learning and its potential impact on various machine learning applications. We believe that this technical overview will contribute to the advancement of meta-learning and its practical implications in addressing realworld problems.

Place, publisher, year, edition, pages
Piscataway, NJ: IEEE, 2024
Keywords
Adaptation models, Data models, deep neural networks, few-shot learning, Meta-learning, Metalearning, representation learning, Surveys, Task analysis, Training, transfer learning, Transfer learning
National Category
Robotics
Identifiers
urn:nbn:se:hh:diva-52730 (URN)10.1109/TPAMI.2024.3357847 (DOI)2-s2.0-85183973598 (Scopus ID)
Funder
Knowledge Foundation
Available from: 2024-02-23 Created: 2024-02-23 Last updated: 2025-01-07Bibliographically approved
2. Meta-learning for efficient unsupervised domain adaptation
Open this publication in new window or tab >>Meta-learning for efficient unsupervised domain adaptation
2024 (English)In: Neurocomputing, ISSN 0925-2312, E-ISSN 1872-8286, Vol. 574, article id 127264Article in journal (Refereed) Published
Abstract [en]

The standard machine learning assumption that training and test data are drawn from the same probability distribution does not hold in many real-world applications due to the inability to reproduce testing conditions at training time. Existing unsupervised domain adaption (UDA) methods address this problem by learning a domain-invariant feature space that performs well on available source domain(s) (labeled training data) and the specific target domain (unlabeled test data). In contrast, instead of simply adapting to domains, this paper aims for an approach that learns to adapt effectively to new unlabeled domains. To do so, we leverage meta-learning to optimize a neural network such that an unlabeled adaptation of its parameters to any domain would yield a good generalization on this latter. The experimental evaluation shows that the proposed approach outperforms standard approaches even when a small amount of unlabeled test data is used for adaptation, demonstrating the benefit of meta-learning prior knowledge from various domains to solve UDA problems.

Place, publisher, year, edition, pages
Amsterdam: Elsevier, 2024
Keywords
Domain adaptation, Meta-learning, Unsupervised learning, Distribution shift
National Category
Computer Sciences
Identifiers
urn:nbn:se:hh:diva-52450 (URN)10.1016/j.neucom.2024.127264 (DOI)001170864800001 ()2-s2.0-85184141702 (Scopus ID)
Funder
Knowledge Foundation
Available from: 2024-01-22 Created: 2024-01-22 Last updated: 2025-01-07Bibliographically approved
3. Meta-Learning from Multimodal Task Distributions Using Multiple Sets of Meta-Parameters
Open this publication in new window or tab >>Meta-Learning from Multimodal Task Distributions Using Multiple Sets of Meta-Parameters
2023 (English)In: 2023 International Joint Conference on Neural Networks (IJCNN), Piscataway, NJ: IEEE, 2023, p. 1-8Conference paper, Published paper (Refereed)
Abstract [en]

Meta-learning or learning to learn involves training a model on various learning tasks in a way that allows it to quickly learn new tasks from the same distribution using only a small amount of training data (i.e., few-shot learning). Current meta-learning methods implicitly assume that the distribution over tasks is unimodal and consists of tasks belonging to a common domain, which significantly reduces the variety of task distributions they can handle. However, in real-world applications, tasks are often very diverse and come from multiple different domains, making it challenging to meta-learn common knowledge shared across the entire task distribution. In this paper, we propose a method for meta-learning from a multimodal task distribution. The proposed method learns multiple sets of meta-parameters (acting as different initializations of a neural network model) and uses a task encoder to select the best initialization to fine-tune for a new task. More specifically, with a few training examples from a task sampled from an unknown mode, the proposed method predicts which set of meta-parameters (i.e., model’s initialization) would lead to a fast adaptation and a good post-adaptation performance on that task. We evaluate the proposed method on a diverse set of few-shot regression and image classification tasks. The results demonstrate the superiority of the proposed method compared to other state of-the-art meta-learning methods and the benefit of learning multiple model initializations when tasks are sampled from a multimodal task distribution. © 2023 IEEE.

Place, publisher, year, edition, pages
Piscataway, NJ: IEEE, 2023
Keywords
Meta-Learning, Few-Shot Learning, Transfer Learning, Task Representation, Multimodal Distribution
National Category
Computer Sciences
Identifiers
urn:nbn:se:hh:diva-51352 (URN)10.1109/IJCNN54540.2023.10191944 (DOI)001046198707013 ()2-s2.0-85169561819 (Scopus ID)978-1-6654-8867-9 (ISBN)
Conference
International Joint Conference on Neural Networks (IJCNN 2023), Gold Coast, Australia, 18-23 June, 2023
Available from: 2023-08-10 Created: 2023-08-10 Last updated: 2025-01-07Bibliographically approved
4. Multimodal meta-learning through meta-learned task representations
Open this publication in new window or tab >>Multimodal meta-learning through meta-learned task representations
2024 (English)In: Neural Computing & Applications, ISSN 0941-0643, E-ISSN 1433-3058, Vol. 36, no 15, p. 8519-8529Article in journal (Refereed) Published
Abstract [en]

Few-shot meta-learning involves training a model on multiple tasks to enable it to efficiently adapt to new, previously unseen tasks with only a limited number of samples. However, current meta-learning methods assume that all tasks are closely related and belong to a common domain, whereas in practice, tasks can be highly diverse and originate from multiple domains, resulting in a multimodal task distribution. This poses a challenge for existing methods as they struggle to learn a shared representation that can be easily adapted to all tasks within the distribution. To address this challenge, we propose a meta-learning framework that can handle multimodal task distributions by conditioning the model on the current task, resulting in a faster adaptation. Our proposed method learns to encode each task and generate task embeddings that modulate the model’s activations. The resulting modulated model becomes specialized for the current task and leads to more effective adaptation. Our framework is designed to work in a realistic setting where the mode from which a task is sampled is unknown. Nonetheless, we also explore the possibility of incorporating auxiliary information, such as the task-mode-label, to further enhance the performance of our method if such information is available. We evaluate our proposed framework on various few-shot regression and image classification tasks, demonstrating its superiority over other state-of-the-art meta-learning methods. The results highlight the benefits of learning to embed task-specific information in the model to guide the adaptation when tasks are sampled from a multimodal distribution. © The Author(s) 2024.

Place, publisher, year, edition, pages
London: Springer, 2024
Keywords
Meta-learning, Few-shot learning, Transfer learning, Task representation, Multimodal distribution
National Category
Computer Sciences
Identifiers
urn:nbn:se:hh:diva-52924 (URN)10.1007/s00521-024-09540-1 (DOI)001167510900006 ()2-s2.0-85187241643 (Scopus ID)
Funder
Halmstad UniversityKnowledge Foundation
Available from: 2024-03-21 Created: 2024-03-21 Last updated: 2025-01-07Bibliographically approved
5. Personalized Federated Learning with Contextual Modulation and Meta-Learning
Open this publication in new window or tab >>Personalized Federated Learning with Contextual Modulation and Meta-Learning
2024 (English)In: Proceedings of the 2024 SIAM International Conference on Data Mining (SDM) / [ed] Shashi Shekhar; Vagelis Papalexakis; Jing Gao; Zhe Jiang; Matteo Riondato, Philadelphia, PA: Society for Industrial and Applied Mathematics, 2024, p. 842-850Conference paper, Published paper (Refereed)
Abstract [en]

Federated learning has emerged as a promising approach for training machine learning models on decentralized data sources while preserving data privacy. However, challenges such as communication bottlenecks, heterogeneity of client devices, and non-i.i.d. data distribution pose significant obstacles to achieving optimal model performance. We propose a novel framework that combines federated learning with meta-learning techniques to enhance both efficiency and generalization capabilities. Our approach introduces a federated modulator that learns contextual information from data batches and uses this knowledge to generate modulation parameters. These parameters dynamically adjust the activations of a base model, which operates using a MAML-based approach for model personalization. Experimental results across diverse datasets highlight the improvements in convergence speed and model performance compared to existing federated learning approaches. These findings highlight the potential of incorporating contextual information and meta-learning techniques into federated learning, paving the way for advancements in distributed machine learning paradigms. Copyright © 2024 by SIAM.

Place, publisher, year, edition, pages
Philadelphia, PA: Society for Industrial and Applied Mathematics, 2024
Series
Proceedings of the ... SIAM International Conference on Data Mining, ISSN 2167-0102, E-ISSN 2167-0099
Keywords
Personalized federated learning, Meta-learning, Federated learning, Context learning
National Category
Computer Sciences
Identifiers
urn:nbn:se:hh:diva-53426 (URN)10.1137/1.9781611978032.96 (DOI)2-s2.0-85193511974 (Scopus ID)978-1-61197-803-2 (ISBN)
Conference
SIAM International Conference on Data Mining (SDM), Houston, USA, April 18-20, 2024
Funder
Knowledge Foundation
Available from: 2024-05-28 Created: 2024-05-28 Last updated: 2025-01-07Bibliographically approved
6. Learning to Learn without Forgetting using Attention
Open this publication in new window or tab >>Learning to Learn without Forgetting using Attention
2024 (English)Conference paper, Published paper (Refereed)
Abstract [en]

Continual learning (CL) refers to the ability to continually learn over time by accommodating new knowledge while retaining previously learned experience. While this concept is inherent in human learning, current machine learning methods are highly prone to overwrite previously learned patterns and thus forget past experience. Instead, model parameters should be updated selectively and carefully, avoiding unnecessary forgetting while optimally leveraging previously learned patterns to accelerate future learning. Since hand-crafting effective update mechanisms is difficult, we propose meta-learning a transformer-based optimizer to enhance CL. This meta-learned optimizer uses attention to learn the complex relationships between model parameters across a stream of tasks, and is designed to generate effective weight updates for the current task while preventing catastrophic forgetting on previously encountered tasks. Evaluations on benchmark datasets like SplitMNIST, RotatedMNIST, and SplitCIFAR-100 affirm the efficacy of the proposed approach in terms of both forward and backward transfer, even on small sets of labeled data, highlighting the advantages of integrating a meta-learned optimizer within the continual learning framework. 

Keywords
Continual learning, Meta-learning, Few-shot learning, Catastrophic forgetting
National Category
Computer Sciences
Identifiers
urn:nbn:se:hh:diva-55120 (URN)
Conference
3rd Conference on Lifelong Learning Agents (CoLLAs), Pisa, Italy, July 29 - 1 August, 2024
Available from: 2024-12-19 Created: 2024-12-19 Last updated: 2025-01-07Bibliographically approved
7. Efficient Few-Shot Human Activity Recognition Via Meta-Learning and Data Augmentation
Open this publication in new window or tab >>Efficient Few-Shot Human Activity Recognition Via Meta-Learning and Data Augmentation
(English)Manuscript (preprint) (Other academic)
Abstract [en]

In the field of Human Activity Recognition (HAR), the rapid evolution of wearable devices necessitates models that are generalizable and can adapt to entirely new subjects and activities with very limited labeled data. Conventional deep learning models, constrained by their reliance on large training datasets and limited adaptability to novel scenarios, face challenges in these settings. This paper introduces a novel few-shot HAR strategy employing meta-learning, which facilitates rapid adaptation to unseen subjects and activities using minimal annotated samples. Our approach augments time series data with a range of transformations, each assigned a learnable weight, enabling the model to prioritize the most effective augmentations and discard the irrelevant ones. Throughout the meta-training phase, the model learns to identify an optimal weighted combination of these transformations, significantly improving the model's adaptability and generalization to new situations with scarce labeled data. During meta-testing, this knowledge enables the model to efficiently learn from and adapt to a very limited set of labeled samples from completely new subjects undertaking entirely new activities. Extensive experiments on various HAR datasets demonstrate our method's enhanced adaptability and generalization to tasks never encountered during training, affirming its potential for real-world applications characterized by limited data availability.

Keywords
Meta-learning, Data augmentation, Time series data, Few-shot learning, Human activity recognition
National Category
Computer Sciences
Identifiers
urn:nbn:se:hh:diva-55121 (URN)
Available from: 2024-12-19 Created: 2024-12-19 Last updated: 2025-01-07Bibliographically approved
8. Unsupervised Meta-Learning via In-Context Learning
Open this publication in new window or tab >>Unsupervised Meta-Learning via In-Context Learning
(English)Manuscript (preprint) (Other academic)
Abstract [en]

Unsupervised meta-learning aims to learn feature representations from unsupervised datasets that can transfer to downstream tasks with limited labeled data. In this paper, we propose a novel approach to unsupervised meta-learning that leverages the generalization abilities of in-context learning observed in transformer architectures. Our method reframes meta-learning as a sequence modeling problem, enabling the transformer encoder to learn task context from support images and utilize it to predict query images. At the core of our approach lies the creation of diverse tasks generated using a combination of data augmentations and a mixing strategy that challenges the model during training while fostering generalization to unseen tasks at test time. Experimental results on benchmark datasets showcase the superiority of our approach over existing unsupervised meta-learning baselines, establishing it as the new state-of-the-art in the field. Remarkably, our method achieves competitive results with supervised and self-supervised approaches, underscoring the efficacy of the model in leveraging generalization over memorization.

National Category
Computer Sciences
Identifiers
urn:nbn:se:hh:diva-55122 (URN)
Available from: 2024-12-19 Created: 2024-12-19 Last updated: 2025-01-07Bibliographically approved

Open Access in DiVA

fulltext(21867 kB)69 downloads
File information
File name FULLTEXT01.pdfFile size 21867 kBChecksum SHA-512
a3c80745e3e14c418d9917f2b604f43452336f43f03aa1e46fc2ea4e793b163b4155619b8d4ae4ee09d45218817e4a309749cbd0e634b4515c2cbd2923d53b9e
Type fulltextMimetype application/pdf

Authority records

Vettoruzzo, Anna

Search in DiVA

By author/editor
Vettoruzzo, Anna
By organisation
School of Information Technology
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 69 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

isbn
urn-nbn

Altmetric score

isbn
urn-nbn
Total: 198 hits
12 2 of 2
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf