Ammonia-Net: A Multi-task Joint Learning Model for Multi-class
Segmentation and Classification in Tooth-marked Tongue Diagnosis
- URL: http://arxiv.org/abs/2310.03472v1
- Date: Thu, 5 Oct 2023 11:28:32 GMT
- Title: Ammonia-Net: A Multi-task Joint Learning Model for Multi-class
Segmentation and Classification in Tooth-marked Tongue Diagnosis
- Authors: Shunkai Shi, Yuqi Wang, Qihui Ye, Yanran Wang, Yiming Zhu, Muhammad
Hassan, Aikaterini Melliou, Dongmei Yu
- Abstract summary: In Traditional Chinese Medicine, the tooth marks on the tongue serve as a crucial indicator for assessing qi (yang) deficiency.
To address these problems, we propose a multi-task joint learning model named Ammonia-Net.
Ammonia-Net performs semantic segmentation of tongue images to identify tongue and tooth marks.
- Score: 12.095100353695038
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In Traditional Chinese Medicine, the tooth marks on the tongue, stemming from
prolonged dental pressure, serve as a crucial indicator for assessing qi (yang)
deficiency, which is intrinsically linked to visceral health. Manual diagnosis
of tooth-marked tongue solely relies on experience. Nonetheless, the diversity
in shape, color, and type of tooth marks poses a challenge to diagnostic
accuracy and consistency. To address these problems, herein we propose a
multi-task joint learning model named Ammonia-Net. This model employs a
convolutional neural network-based architecture, specifically designed for
multi-class segmentation and classification of tongue images. Ammonia-Net
performs semantic segmentation of tongue images to identify tongue and tooth
marks. With the assistance of segmentation output, it classifies the images
into the desired number of classes: healthy tongue, light tongue, moderate
tongue, and severe tongue. As far as we know, this is the first attempt to
apply the semantic segmentation results of tooth marks for tooth-marked tongue
classification. To train Ammonia-Net, we collect 856 tongue images from 856
subjects. After a number of extensive experiments, the experimental results
show that the proposed model achieves 99.06% accuracy in the two-class
classification task of tooth-marked tongue identification and 80.02%. As for
the segmentation task, mIoU for tongue and tooth marks amounts to 71.65%.
Related papers
- Weakly Supervised Object Detection for Automatic Tooth-marked Tongue Recognition [19.34036038278796]
Tongue diagnosis in Traditional Chinese Medicine (TCM) is a crucial diagnostic method that can reflect an individual's health status.
Traditional methods for identifying tooth-marked tongues are subjective and inconsistent because they rely on practitioner experience.
We propose a novel fully automated WeaklySupervised method using Vision transformer and Multiple instance learning WSVM for tongue extraction and tooth-marked tongue recognition.
arXiv Detail & Related papers (2024-08-29T11:31:28Z) - TongueSAM: An Universal Tongue Segmentation Model Based on SAM with
Zero-Shot [8.211898244288305]
Tongue segmentation serves as the primary step in automated TCM tongue diagnosis.
This paper proposes a universal tongue segmentation model named TongueSAM based on SAM (Segment Anything Model)
arXiv Detail & Related papers (2023-08-12T02:38:43Z) - Multitasking Models are Robust to Structural Failure: A Neural Model for
Bilingual Cognitive Reserve [78.3500985535601]
We find a surprising connection between multitask learning and robustness to neuron failures.
Our experiments show that bilingual language models retain higher performance under various neuron perturbations.
We provide a theoretical justification for this robustness by mathematically analyzing linear representation learning.
arXiv Detail & Related papers (2022-10-20T22:23:27Z) - Tooth Instance Segmentation on Panoramic Dental Radiographs Using U-Nets
and Morphological Processing [0.0]
We propose a post-processing stage to obtain a segmentation map in which the objects in the image are separated.
The proposed post-processing stages reduce the mean error of tooth count to 6.15%, whereas the error without post-processing is 26.81%.
arXiv Detail & Related papers (2022-03-31T21:11:51Z) - Interpretable and Interactive Deep Multiple Instance Learning for Dental
Caries Classification in Bitewing X-rays [6.98282810587309]
We propose a simple and efficient image classification architecture based on deep multiple instance learning.
Our approach contributes in two ways: First, it outputs a heatmap of local patch classification probabilities despite being trained with weak image-level labels.
In contrast to existing methods, the human user can faithfully interpret predictions and interact with the model to decide which regions to attend to.
arXiv Detail & Related papers (2021-12-17T14:09:27Z) - Developing a Novel Approach for Periapical Dental Radiographs
Segmentation [1.332560004325655]
The proposed algorithm is made of two stages. The first stage is pre-processing.
The second and main part of this algorithm calculated rotation degree and uses the integral projection method for tooth isolation.
Experimental results show that this algorithm is robust and achieves high accuracy.
arXiv Detail & Related papers (2021-11-13T17:25:35Z) - Two-Stage Mesh Deep Learning for Automated Tooth Segmentation and
Landmark Localization on 3D Intraoral Scans [56.55092443401416]
emphiMeshSegNet in the first stage of TS-MDL reached an averaged Dice similarity coefficient (DSC) at 0.953pm0.076$, significantly outperforming the original MeshSegNet.
PointNet-Reg achieved a mean absolute error (MAE) of $0.623pm0.718, mm$ in distances between the prediction and ground truth for $44$ landmarks, which is superior compared with other networks for landmark detection.
arXiv Detail & Related papers (2021-09-24T13:00:26Z) - CBLUE: A Chinese Biomedical Language Understanding Evaluation Benchmark [51.38557174322772]
We present the first Chinese Biomedical Language Understanding Evaluation benchmark.
It is a collection of natural language understanding tasks including named entity recognition, information extraction, clinical diagnosis normalization, single-sentence/sentence-pair classification.
We report empirical results with the current 11 pre-trained Chinese models, and experimental results show that state-of-the-art neural models perform by far worse than the human ceiling.
arXiv Detail & Related papers (2021-06-15T12:25:30Z) - AM2iCo: Evaluating Word Meaning in Context across Low-ResourceLanguages
with Adversarial Examples [51.048234591165155]
We present AM2iCo, Adversarial and Multilingual Meaning in Context.
It aims to faithfully assess the ability of state-of-the-art (SotA) representation models to understand the identity of word meaning in cross-lingual contexts.
Results reveal that current SotA pretrained encoders substantially lag behind human performance.
arXiv Detail & Related papers (2021-04-17T20:23:45Z) - Leveraging Adversarial Training in Self-Learning for Cross-Lingual Text
Classification [52.69730591919885]
We present a semi-supervised adversarial training process that minimizes the maximal loss for label-preserving input perturbations.
We observe significant gains in effectiveness on document and intent classification for a diverse set of languages.
arXiv Detail & Related papers (2020-07-29T19:38:35Z) - Unsupervised Cross-lingual Representation Learning for Speech
Recognition [63.85924123692923]
XLSR learns cross-lingual speech representations by pretraining a single model from the raw waveform of speech in multiple languages.
We build on wav2vec 2.0 which is trained by solving a contrastive task over masked latent speech representations.
Experiments show that cross-lingual pretraining significantly outperforms monolingual pretraining.
arXiv Detail & Related papers (2020-06-24T18:25:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.