Skip to content
/ MLTrMR Public

Code and dataset for paper "Masked Latent Transformer with the Random Masking Ratio to Advance the Diagnosis of Dental Fluorosis."

Notifications You must be signed in to change notification settings

uxhao-o/MLTrMR

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 

Repository files navigation

MLTrMR

Masked Latent Transformer with the Random Masking Ratio to Advance the Diagnosis of Dental Fluorosis

Hao Xu1,2, Yun Wu1,2, 📧, Junpeng Wu3, Rui Xie3, Maohua Gu1,2, Rongpin Wang4

1 State Key Laboratory of Public Big Data, Guizhou University
2 College of Computer Science and Technology, Guizhou University
3 Zhijin County People’s Hospital
4 Department of Medical Imaging, Guizhou Provincial People’s Hospital

( 📧 ) Corresponding author.

ArXiv Preprint (arXiv 2404.13564)

Notes

  • This study has been submitted to journal for review.
  • We will publish the code and dataset of MLTrMR in this repository when our paper is accepted.

Abstract

Dental fluorosis is a chronic disease caused by long-term overconsumption of fluoride, which leads to changes in the appearance of tooth enamel. It is an important basis for early non-invasive diagnosis of endemic fluorosis. However, even dental professionals may not be able to accurately distinguish dental fluorosis and its severity based on tooth images. Currently, there is limited research on applying deep learning to diagnosing dental fluorosis. Therefore, we propose a novel deep learning model called masked latent transformer with the random masking ratio (MLTrMR). MLTrMR introduces a mask latent modeling scheme based on Vision Transformer to enhance contextual learning of dental fluorosis lesion characteristics. Consisting of a latent embedder, encoder, and decoder, MLTrMR employs the latent embedder to extract latent tokens from the original image, whereas the encoder and decoder comprising the latent transformer (LT) block are used to process unmasked tokens and predict masked tokens, respectively. To mitigate the lack of inductive bias in Vision Transformer, which may result in performance degradation, the LT block introduces latent tokens to enhance the learning capacity of latent lesion features. Furthermore, we design an auxiliary loss function to constrain the parameter update direction of the model. MLTrMR achieves 80.19% accuracy, 75.79% F1, and 81.28% quadratic weighted kappa on we construct the first open-source dental fluorosis image dataset (DFID), making it state-of-the-art (SOTA).

Overview

Dental Fluorosis Image Dataset (DFID)

DFID

About

Code and dataset for paper "Masked Latent Transformer with the Random Masking Ratio to Advance the Diagnosis of Dental Fluorosis."

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published