Spatial machine-learning model diagnostics: a model-agnostic
distance-based approach
- URL: http://arxiv.org/abs/2111.08478v1
- Date: Sat, 13 Nov 2021 01:50:36 GMT
- Title: Spatial machine-learning model diagnostics: a model-agnostic
distance-based approach
- Authors: Alexander Brenning
- Abstract summary: This contribution proposes spatial prediction error profiles (SPEPs) and spatial variable importance profiles (SVIPs) as novel model-agnostic assessment and interpretation tools.
The SPEPs and SVIPs of geostatistical methods, linear models, random forest, and hybrid algorithms show striking differences and also relevant similarities.
The novel diagnostic tools enrich the toolkit of spatial data science, and may improve ML model interpretation, selection, and design.
- Score: 91.62936410696409
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: While significant progress has been made towards explaining black-box
machine-learning (ML) models, there is still a distinct lack of diagnostic
tools that elucidate the spatial behaviour of ML models in terms of predictive
skill and variable importance. This contribution proposes spatial prediction
error profiles (SPEPs) and spatial variable importance profiles (SVIPs) as
novel model-agnostic assessment and interpretation tools for spatial prediction
models with a focus on prediction distance. Their suitability is demonstrated
in two case studies representing a regionalization task in an
environmental-science context, and a classification task from remotely-sensed
land cover classification. In these case studies, the SPEPs and SVIPs of
geostatistical methods, linear models, random forest, and hybrid algorithms
show striking differences but also relevant similarities. Limitations of
related cross-validation techniques are outlined, and the case is made that
modelers should focus their model assessment and interpretation on the intended
spatial prediction horizon. The range of autocorrelation, in contrast, is not a
suitable criterion for defining spatial cross-validation test sets. The novel
diagnostic tools enrich the toolkit of spatial data science, and may improve ML
model interpretation, selection, and design.
Related papers
- Multi-Scale and Multimodal Species Distribution Modeling [4.022195138381868]
Species distribution models (SDMs) aim to predict the distribution of species relating occurrence data with environmental variables.
Recent applications of deep learning to SDMs have enabled new avenues, specifically the inclusion of spatial data.
We develop a modular structure for SDMs that allows us to test the effect of scale in both single- and multi-scale settings.
Results on the GeoLifeCLEF 2023 benchmark indicate that considering multimodal data and learning multi-scale representations leads to more accurate models.
arXiv Detail & Related papers (2024-11-06T15:57:20Z) - On conditional diffusion models for PDE simulations [53.01911265639582]
We study score-based diffusion models for forecasting and assimilation of sparse observations.
We propose an autoregressive sampling approach that significantly improves performance in forecasting.
We also propose a new training strategy for conditional score-based models that achieves stable performance over a range of history lengths.
arXiv Detail & Related papers (2024-10-21T18:31:04Z) - Influence Functions for Scalable Data Attribution in Diffusion Models [52.92223039302037]
Diffusion models have led to significant advancements in generative modelling.
Yet their widespread adoption poses challenges regarding data attribution and interpretability.
In this paper, we aim to help address such challenges by developing an textitinfluence functions framework.
arXiv Detail & Related papers (2024-10-17T17:59:02Z) - Determining Domain of Machine Learning Models using Kernel Density Estimates: Applications in Materials Property Prediction [1.8551396341435895]
We develop a new approach of assessing model domain using kernel density estimation.
We show that chemical groups considered unrelated based on established chemical knowledge exhibit significant dissimilarities by our measure.
High measures of dissimilarity are associated with poor model performance and poor estimates of model uncertainty.
arXiv Detail & Related papers (2024-05-28T15:41:16Z) - The CAST package for training and assessment of spatial prediction models in R [0.0]
We introduce the CAST package and its core functionalities.
We will go through the different steps of the modelling workflow and show how CAST can be used to support more reliable spatial predictions.
arXiv Detail & Related papers (2024-04-10T12:48:10Z) - WEEP: A method for spatial interpretation of weakly supervised CNN models in computational pathology [0.36096289461554343]
We propose a novel method, Wsi rEgion sElection aPproach (WEEP), for model interpretation.
We demonstrate WEEP on a binary classification task in the area of breast cancer computational pathology.
arXiv Detail & Related papers (2024-03-22T14:32:02Z) - A comprehensive comparative evaluation and analysis of Distributional
Semantic Models [61.41800660636555]
We perform a comprehensive evaluation of type distributional vectors, either produced by static DSMs or obtained by averaging the contextualized vectors generated by BERT.
The results show that the alleged superiority of predict based models is more apparent than real, and surely not ubiquitous.
We borrow from cognitive neuroscience the methodology of Representational Similarity Analysis (RSA) to inspect the semantic spaces generated by distributional models.
arXiv Detail & Related papers (2021-05-20T15:18:06Z) - Transforming Feature Space to Interpret Machine Learning Models [91.62936410696409]
This contribution proposes a novel approach that interprets machine-learning models through the lens of feature space transformations.
It can be used to enhance unconditional as well as conditional post-hoc diagnostic tools.
A case study on remote-sensing landcover classification with 46 features is used to demonstrate the potential of the proposed approach.
arXiv Detail & Related papers (2021-04-09T10:48:11Z) - Models, Pixels, and Rewards: Evaluating Design Trade-offs in Visual
Model-Based Reinforcement Learning [109.74041512359476]
We study a number of design decisions for the predictive model in visual MBRL algorithms.
We find that a range of design decisions that are often considered crucial, such as the use of latent spaces, have little effect on task performance.
We show how this phenomenon is related to exploration and how some of the lower-scoring models on standard benchmarks will perform the same as the best-performing models when trained on the same training data.
arXiv Detail & Related papers (2020-12-08T18:03:21Z) - Predicting into unknown space? Estimating the area of applicability of
spatial prediction models [0.0]
We suggest a methodology that delineates the "area of applicability" (AOA) that we define as the area, for which the cross-validation error of the model applies.
We test for the ideal threshold by using simulated data and compare the prediction error within the AOA with the cross-validation error of the model.
arXiv Detail & Related papers (2020-05-16T10:31:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.