Meta-learning an Intermediate Representation for Few-shot Block-wise
Prediction of Landslide Susceptibility
- URL: http://arxiv.org/abs/2110.04922v1
- Date: Sun, 3 Oct 2021 05:40:50 GMT
- Title: Meta-learning an Intermediate Representation for Few-shot Block-wise
Prediction of Landslide Susceptibility
- Authors: Li Chen, Yulin Ding, Han Hu, Qing Zhu, Haowei Zeng, Haojia Yu, Qisen
Shang, Yongfei Song
- Abstract summary: Current methods generally apply a single global model to predict the landslide susceptibility map (LSM) for an entire target region.
We argue that, in complex circumstances, each part of the region holds different landslide-inducing environments, and therefore, should be predicted individually with respective models.
We train an intermediate representation by the meta-learning paradigm, which is superior for capturing information from LSM tasks in order to generalize proficiently.
- Score: 12.299036699550374
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Predicting a landslide susceptibility map (LSM) is essential for risk
recognition and disaster prevention. Despite the successful application of
data-driven prediction approaches, current data-driven methods generally apply
a single global model to predict the LSM for an entire target region. However,
we argue that, in complex circumstances, especially in large-scale areas, each
part of the region holds different landslide-inducing environments, and
therefore, should be predicted individually with respective models. In this
study, target scenarios were segmented into blocks for individual analysis
using topographical factors. But simply conducting training and testing using
limited samples within each block is hardly possible for a satisfactory LSM
prediction, due to the adverse effect of \textit{overfitting}. To solve the
problems, we train an intermediate representation by the meta-learning
paradigm, which is superior for capturing information from LSM tasks in order
to generalize proficiently. We chose this based on the hypothesis that there
are more general concepts among LSM tasks that are sensitive to variations in
input features. Thus, using the intermediate representation, we can easily
adapt the model for different blocks or even unseen tasks using few exemplar
samples. Experimental results on two study areas demonstrated the validity of
our block-wise analysis in large scenarios and revealed the top few-shot
adaption performances of the proposed methods.
Related papers
- MIRRAMS: Towards Training Models Robust to Missingness Distribution Shifts [2.5357049657770516]
In real-world data analysis, missingness distributional shifts between training and test input datasets frequently occur.<n>We propose a novel deep learning framework designed to address such shifts in missingness distributions.<n>Our approach achieves state-of-the-art performance even without missing data and can be naturally extended to address semi-supervised learning tasks.
arXiv Detail & Related papers (2025-07-11T03:03:30Z) - Can Large Language Models Trigger a Paradigm Shift in Travel Behavior Modeling? Experiences with Modeling Travel Satisfaction [2.2974830861901414]
This study uses data on travel satisfaction from a household survey in shanghai to identify the existence and source of misalignment between Large Language Models and human behavior.<n>We find that the zero-shot LLM exhibits behavioral misalignment, resulting in relatively low prediction accuracy.<n>We propose an LLM-based modeling approach that can be applied to model travel behavior using samples of small sizes.
arXiv Detail & Related papers (2025-05-29T09:11:58Z) - Generalized Semantic Contrastive Learning via Embedding Side Information for Few-Shot Object Detection [52.490375806093745]
The objective of few-shot object detection (FSOD) is to detect novel objects with few training samples.
We introduce the side information to alleviate the negative influences derived from the feature space and sample viewpoints.
Our model outperforms the previous state-of-the-art methods, significantly improving the ability of FSOD in most shots/splits.
arXiv Detail & Related papers (2025-04-09T17:24:05Z) - Towards Understanding Extrapolation: a Causal Lens [53.15488984371969]
We provide a theoretical understanding of when extrapolation is possible and offer principled methods to achieve it.
Under this formulation, we cast the extrapolation problem into a latent-variable identification problem.
Our theory reveals the intricate interplay between the underlying manifold's smoothness and the shift properties.
arXiv Detail & Related papers (2025-01-15T21:29:29Z) - Learning from Different Samples: A Source-free Framework for Semi-supervised Domain Adaptation [20.172605920901777]
This paper focuses on designing a framework to use different strategies for comprehensively mining different target samples.
We propose a novel source-free framework (SOUF) to achieve semi-supervised fine-tuning of the source pre-trained model on the target domain.
arXiv Detail & Related papers (2024-11-11T02:09:32Z) - Reducing Spurious Correlation for Federated Domain Generalization [15.864230656989854]
In open-world scenarios, global models may struggle to predict well on entirely new domain data captured by certain media.
Existing methods still rely on strong statistical correlations between samples and labels to address this issue.
We introduce FedCD, an overall optimization framework at both the local and global levels.
arXiv Detail & Related papers (2024-07-27T05:06:31Z) - Supervised Fine-Tuning as Inverse Reinforcement Learning [8.044033685073003]
The prevailing approach to aligning Large Language Models (LLMs) typically relies on human or AI feedback.
In our work, we question the efficacy of such datasets and explore various scenarios where alignment with expert demonstrations proves more realistic.
arXiv Detail & Related papers (2024-03-18T17:52:57Z) - Source-Free Unsupervised Domain Adaptation with Hypothesis Consolidation
of Prediction Rationale [53.152460508207184]
Source-Free Unsupervised Domain Adaptation (SFUDA) is a challenging task where a model needs to be adapted to a new domain without access to target domain labels or source domain data.
This paper proposes a novel approach that considers multiple prediction hypotheses for each sample and investigates the rationale behind each hypothesis.
To achieve the optimal performance, we propose a three-step adaptation process: model pre-adaptation, hypothesis consolidation, and semi-supervised learning.
arXiv Detail & Related papers (2024-02-02T05:53:22Z) - Querying Easily Flip-flopped Samples for Deep Active Learning [63.62397322172216]
Active learning is a machine learning paradigm that aims to improve the performance of a model by strategically selecting and querying unlabeled data.
One effective selection strategy is to base it on the model's predictive uncertainty, which can be interpreted as a measure of how informative a sample is.
This paper proposes the it least disagree metric (LDM) as the smallest probability of disagreement of the predicted label.
arXiv Detail & Related papers (2024-01-18T08:12:23Z) - An Additive Instance-Wise Approach to Multi-class Model Interpretation [53.87578024052922]
Interpretable machine learning offers insights into what factors drive a certain prediction of a black-box system.
Existing methods mainly focus on selecting explanatory input features, which follow either locally additive or instance-wise approaches.
This work exploits the strengths of both methods and proposes a global framework for learning local explanations simultaneously for multiple target classes.
arXiv Detail & Related papers (2022-07-07T06:50:27Z) - Learning Invariant Representation with Consistency and Diversity for
Semi-supervised Source Hypothesis Transfer [46.68586555288172]
We propose a novel task named Semi-supervised Source Hypothesis Transfer (SSHT), which performs domain adaptation based on source trained model, to generalize well in target domain with a few supervisions.
We propose Consistency and Diversity Learning (CDL), a simple but effective framework for SSHT by facilitating prediction consistency between two randomly augmented unlabeled data.
Experimental results show that our method outperforms existing SSDA methods and unsupervised model adaptation methods on DomainNet, Office-Home and Office-31 datasets.
arXiv Detail & Related papers (2021-07-07T04:14:24Z) - Counterfactual Maximum Likelihood Estimation for Training Deep Networks [83.44219640437657]
Deep learning models are prone to learning spurious correlations that should not be learned as predictive clues.
We propose a causality-based training framework to reduce the spurious correlations caused by observable confounders.
We conduct experiments on two real-world tasks: Natural Language Inference (NLI) and Image Captioning.
arXiv Detail & Related papers (2021-06-07T17:47:16Z) - Learning Invariant Representations and Risks for Semi-supervised Domain
Adaptation [109.73983088432364]
We propose the first method that aims to simultaneously learn invariant representations and risks under the setting of semi-supervised domain adaptation (Semi-DA)
We introduce the LIRR algorithm for jointly textbfLearning textbfInvariant textbfRepresentations and textbfRisks.
arXiv Detail & Related papers (2020-10-09T15:42:35Z) - Learning Diverse Representations for Fast Adaptation to Distribution
Shift [78.83747601814669]
We present a method for learning multiple models, incorporating an objective that pressures each to learn a distinct way to solve the task.
We demonstrate our framework's ability to facilitate rapid adaptation to distribution shift.
arXiv Detail & Related papers (2020-06-12T12:23:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.