MISeval: a Metric Library for Medical Image Segmentation Evaluation
- URL: http://arxiv.org/abs/2201.09395v1
- Date: Sun, 23 Jan 2022 23:06:47 GMT
- Title: MISeval: a Metric Library for Medical Image Segmentation Evaluation
- Authors: Dominik M\"uller, Dennis Hartmann, Philip Meyer, Florian Auer, I\~naki
Soto-Rey and Frank Kramer
- Abstract summary: There is no universal metric library in Python for standardized and reproducible evaluation.
We propose our open-source publicly available Python package MISeval: a metric library for Medical Image Evaluation.
- Score: 1.4680035572775534
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Correct performance assessment is crucial for evaluating modern artificial
intelligence algorithms in medicine like deep-learning based medical image
segmentation models. However, there is no universal metric library in Python
for standardized and reproducible evaluation. Thus, we propose our open-source
publicly available Python package MISeval: a metric library for Medical Image
Segmentation Evaluation. The implemented metrics can be intuitively used and
easily integrated into any performance assessment pipeline. The package
utilizes modern CI/CD strategies to ensure functionality and stability. MISeval
is available from PyPI (miseval) and GitHub:
https://github.com/frankkramer-lab/miseval.
Related papers
- Seg-metrics: a Python package to compute segmentation metrics [0.6827423171182151]
textttseg-metrics is an open-source Python package for standardized MIS model evaluation.
textttseg-metrics supports multiple file formats and is easily installable through the Python Package Index (PyPI)
arXiv Detail & Related papers (2024-01-12T16:30:54Z) - SPRINT: A Unified Toolkit for Evaluating and Demystifying Zero-shot
Neural Sparse Retrieval [92.27387459751309]
We provide SPRINT, a unified Python toolkit for evaluating neural sparse retrieval.
We establish strong and reproducible zero-shot sparse retrieval baselines across the well-acknowledged benchmark, BEIR.
We show that SPLADEv2 produces sparse representations with a majority of tokens outside of the original query and document.
arXiv Detail & Related papers (2023-07-19T22:48:02Z) - Evaluate & Evaluation on the Hub: Better Best Practices for Data and
Model Measurements [167.73134600289603]
evaluate is a library to support best practices for measurements, metrics, and comparisons of data and models.
Evaluation on the Hub is a platform that enables the large-scale evaluation of over 75,000 models and 11,000 datasets.
arXiv Detail & Related papers (2022-09-30T18:35:39Z) - Scikit-dimension: a Python package for intrinsic dimension estimation [58.8599521537]
This technical note introduces textttscikit-dimension, an open-source Python package for intrinsic dimension estimation.
textttscikit-dimension package provides a uniform implementation of most of the known ID estimators based on scikit-learn application programming interface.
We briefly describe the package and demonstrate its use in a large-scale (more than 500 datasets) benchmarking of methods for ID estimation in real-life and synthetic data.
arXiv Detail & Related papers (2021-09-06T16:46:38Z) - PyHealth: A Python Library for Health Predictive Models [53.848478115284195]
PyHealth is an open-source Python toolbox for developing various predictive models on healthcare data.
The data preprocessing module enables the transformation of complex healthcare datasets into machine learning friendly formats.
The predictive modeling module provides more than 30 machine learning models, including established ensemble trees and deep neural network-based approaches.
arXiv Detail & Related papers (2021-01-11T22:02:08Z) - pymia: A Python package for data handling and evaluation in deep
learning-based medical image analysis [0.9176056742068814]
pymia is an open-source Python package for data handling and evaluation in medical image analysis.
The package is highly flexible, allows for fast prototyping, and reduces the burden of implementing data handling routines.
pymia was successfully used in a variety of research projects for segmentation, reconstruction, and regression.
arXiv Detail & Related papers (2020-10-07T20:25:52Z) - Captum: A unified and generic model interpretability library for PyTorch [49.72749684393332]
We introduce a novel, unified, open-source model interpretability library for PyTorch.
The library contains generic implementations of a number of gradient and perturbation-based attribution algorithms.
It can be used for both classification and non-classification models.
arXiv Detail & Related papers (2020-09-16T18:57:57Z) - SacreROUGE: An Open-Source Library for Using and Developing
Summarization Evaluation Metrics [74.28810048824519]
SacreROUGE is an open-source library for using and developing summarization evaluation metrics.
The library provides Python wrappers around the official implementations of existing evaluation metrics.
It provides functionality to evaluate how well any metric implemented in the library correlates to human-annotated judgments.
arXiv Detail & Related papers (2020-07-10T13:26:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.