Parameter Choice and Neuro-Symbolic Approaches for Deep Domain-Invariant Learning
- URL: http://arxiv.org/abs/2410.06235v1
- Date: Tue, 8 Oct 2024 17:46:35 GMT
- Title: Parameter Choice and Neuro-Symbolic Approaches for Deep Domain-Invariant Learning
- Authors: Marius-Constantin Dinu,
- Abstract summary: Neuro-symbolic (NeSy) AI bridges the gap between symbolic and sub-symbolic paradigms.
NeSy AI systems use multiple models and methods to generalize to unseen domains.
This work establishes a framework for scalable and generalizable broad AI systems applicable across various problem settings.
- Score: 2.0547410497538445
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As artificial intelligence (AI) systems advance, we move towards broad AI: systems capable of performing well on diverse tasks, understanding context, and adapting rapidly to new scenarios. A central challenge for broad AI systems is to generalize over tasks in related domains and being robust to distribution shifts. Neuro-symbolic (NeSy) AI bridges the gap between symbolic and sub-symbolic paradigms to address these challenges, enabling adaptable, generalizable, and more interpretable systems. The development of broad AI requires advancements in domain adaptation (DA), enabling models trained on source domains to effectively generalize to unseen target domains. Traditional approaches often rely on parameter optimization and fine-tuning, which can be impractical due to high costs and risks of catastrophic forgetting. NeSy AI systems use multiple models and methods to generalize to unseen domains and maintain performance across varying conditions. We analyze common DA and NeSy approaches with a focus on deep domain-invariant learning, extending to real-world challenges such as adapting to continuously changing domains and handling large domain gaps. We showcase state-of-the-art model-selection methods for scenarios with limited samples and introduce domain-specific adaptations without gradient-based updates for cases where model tuning is infeasible. This work establishes a framework for scalable and generalizable broad AI systems applicable across various problem settings, demonstrating how symbolic reasoning and large language models can build universal computational graphs that generalize across domains and problems, contributing to more adaptable AI approaches for real-world applications.
Related papers
- Learning to Generalize Unseen Domains via Multi-Source Meta Learning for Text Classification [71.08024880298613]
We study the multi-source Domain Generalization of text classification.
We propose a framework to use multiple seen domains to train a model that can achieve high accuracy in an unseen domain.
arXiv Detail & Related papers (2024-09-20T07:46:21Z) - AD-Aligning: Emulating Human-like Generalization for Cognitive Domain Adaptation in Deep Learning [3.3543468626874486]
Domain adaptation is pivotal for enabling deep learning models to generalize across diverse domains.
We introduce AD-Aligning, a novel approach that combines adversarial training with source-target domain alignment.
Our findings highlight AD-Aligning's ability to emulate the nuanced cognitive processes inherent in human perception.
arXiv Detail & Related papers (2024-05-15T02:34:06Z) - Understanding and Estimating Domain Complexity Across Domains [2.1613662656419406]
We propose a general framework for estimating domain complexity across diverse environments.
By analyzing dimensionality, sparsity, and diversity within these categories, we offer a comprehensive view of domain challenges.
arXiv Detail & Related papers (2023-12-20T23:47:17Z) - Design Principles for Model Generalization and Scalable AI Integration
in Radio Access Networks [2.846642778157227]
This paper emphasizes the pivotal role of achieving model generalization in enhancing performance and enabling scalable AI integration within radio communications.
We outline design principles for model generalization in three key domains: environment for robustness, intents for adaptability to system objectives, and control tasks for reducing AI-driven control loops.
We propose a learning architecture that leverages centralization of training and data management functionalities, combined with distributed data generation.
arXiv Detail & Related papers (2023-06-09T20:46:31Z) - Normalization Perturbation: A Simple Domain Generalization Method for
Real-World Domain Shifts [133.99270341855728]
Real-world domain styles can vary substantially due to environment changes and sensor noises.
Deep models only know the training domain style.
We propose Normalization Perturbation to overcome this domain style overfitting problem.
arXiv Detail & Related papers (2022-11-08T17:36:49Z) - Improving Diversity with Adversarially Learned Transformations for
Domain Generalization [81.26960899663601]
We present a novel framework that uses adversarially learned transformations (ALT) using a neural network to model plausible, yet hard image transformations.
We show that ALT can naturally work with existing diversity modules to produce highly distinct, and large transformations of the source domain leading to state-of-the-art performance.
arXiv Detail & Related papers (2022-06-15T18:05:24Z) - Generalizing to Evolving Domains with Latent Structure-Aware Sequential
Autoencoder [32.46804768486719]
We introduce a probabilistic framework called Latent Structure-aware Sequential Autoencoder (LSSAE) to tackle the problem of evolving domain generalization.
Experimental results on both synthetic and real-world datasets show that LSSAE can lead to superior performances.
arXiv Detail & Related papers (2022-05-16T13:11:29Z) - From Big to Small: Adaptive Learning to Partial-Set Domains [94.92635970450578]
Domain adaptation targets at knowledge acquisition and dissemination from a labeled source domain to an unlabeled target domain under distribution shift.
Recent advances show that deep pre-trained models of large scale endow rich knowledge to tackle diverse downstream tasks of small scale.
This paper introduces Partial Domain Adaptation (PDA), a learning paradigm that relaxes the identical class space assumption to that the source class space subsumes the target class space.
arXiv Detail & Related papers (2022-03-14T07:02:45Z) - A Novel Mix-normalization Method for Generalizable Multi-source Person
Re-identification [49.548815417844786]
Person re-identification (Re-ID) has achieved great success in the supervised scenario.
It is difficult to directly transfer the supervised model to arbitrary unseen domains due to the model overfitting to the seen source domains.
We propose MixNorm, which consists of domain-aware mix-normalization (DMN) and domain-ware center regularization (DCR)
arXiv Detail & Related papers (2022-01-24T18:09:38Z) - Efficient Model-Based Multi-Agent Mean-Field Reinforcement Learning [89.31889875864599]
We propose an efficient model-based reinforcement learning algorithm for learning in multi-agent systems.
Our main theoretical contributions are the first general regret bounds for model-based reinforcement learning for MFC.
We provide a practical parametrization of the core optimization problem.
arXiv Detail & Related papers (2021-07-08T18:01:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.