1
|
Şahin B, Eninanç İ. A Deep Learning-Based Approach to Detect Lamina Dura Loss on Periapical Radiographs. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025; 38:545-555. [PMID: 39838226 PMCID: PMC11811344 DOI: 10.1007/s10278-025-01405-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/15/2024] [Revised: 12/19/2024] [Accepted: 01/01/2025] [Indexed: 01/23/2025]
Abstract
This study aimed to develop a custom artificial intelligence (AI) model for detecting lamina dura (LD) loss around the roots of anterior and posterior teeth on intraoral periapical radiographs. A total of 701 periapical radiographs of the anterior and posterior regions retrieved from the Dentomaxillofacial Radiology archives were reviewed. Images were cropped to include only the teeth exhibiting LD loss and those without LD loss, which were labeled as "1" and "0," respectively. The dataset was diversified using image preprocessing and data augmentation techniques. Among the radiographs, 72% were used for training, 18% for validation, and 10% for testing. A custom AI model, consisting of 4 blocks and 49 layers, with a total of 21.2 million parameters, was developed using the TensorFlow library and residual blocks introduced in ResNet architecture. Sensitivity, specificity, accuracy, precision, F1 score, and kappa (κ) coefficients (for intra-observer agreement) were calculated to evaluate the performance of the AI model. When applied to a test set of 71 images, the AI model showed good performance in detecting LD loss, achieving an average sensitivity of 0.730, specificity of 0.706, accuracy of 0.718, precision of 0.730, and an F1 score of 0.730, regardless of the dental region. This study represents the first known application of an AI algorithm tailored to detect LD loss on periapical radiographs. The developed AI model could aid clinicians in making accurate diagnosis and help prevent misdiagnosis.
Collapse
Affiliation(s)
- Büşra Şahin
- Department of Oral and Maxillofacial Radiology, Faculty of Dentistry, Sivas Cumhuriyet University, Sivas, Turkey.
| | - İlknur Eninanç
- Department of Oral and Maxillofacial Radiology, Faculty of Dentistry, Sivas Cumhuriyet University, Sivas, Turkey
| |
Collapse
|
2
|
Güller MT, Kumbasar N, Miloğlu Ö. Evaluation of the effectiveness of panoramic radiography in impacted mandibular third molars on deep learning models developed with findings obtained with cone beam computed tomography. Oral Radiol 2024:10.1007/s11282-024-00799-7. [PMID: 39729224 DOI: 10.1007/s11282-024-00799-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2024] [Accepted: 12/12/2024] [Indexed: 12/28/2024]
Abstract
OBJECTIVE The aim of this study is to determine the contact relationship and position of impacted mandibular third molar teeth (IMM3) with the mandibular canal (MC) in panoramic radiography (PR) images using deep learning (DL) models trained with the help of cone beam computed tomography (CBCT) and DL to compare the performances of the architectures. METHODS In this study, a total of 546 IMM3s from 290 patients with CBCT and PR images were included. The performances of SqueezeNet, GoogLeNet, and Inception-v3 architectures in solving four problems on two different regions of interest (RoI) were evaluated. RESULTS The SqueezeNet architecture performed the best on the vertical RoI, showing 93.2% accuracy in the identification of the 2nd problem (contact relationship buccal or lingual). Inception-v3 showed the highest performance with 84.8% accuracy in horizontal RoI for the 1st problem (contact relationship-no contact relationship), GoogLeNet showed 77.4% accuracy in horizontal RoI for the 4th problem (contact relationship buccal, lingual, other category, or no contact relationship), and GoogLeNet showed 70.0% accuracy in horizontal RoI for the 3rd problem (contact relationship buccal, lingual, or other category). CONCLUSION This study found that the Inception-v3 model showed the highest accuracy values in determining the contact relationship, and SqueezeNet architecture showed the highest accuracy values in determining the position of IMM3 relative to MC in the presence of a contact relationship.
Collapse
Affiliation(s)
- Mustafa Taha Güller
- Department of Oral, Dental and Maxillofacial Radiology, Faculty of Dentistry, Giresun University, Giresun, 28200, Turkey
| | - Nida Kumbasar
- TUBITAK, Informatics and Information Security Research Center (BILGEM), Kocaeli, 41470, Turkey
- Department of Electrical Electronic Engineering, Faculty of Engineering, Ataturk University, Erzurum, 25240, Turkey
| | - Özkan Miloğlu
- Department of Oral, Dental and Maxillofacial Radiology, Faculty of Dentistry, Ataturk University, Erzurum, 25240, Turkey.
| |
Collapse
|
3
|
Ünal SY, Namdar Pekiner F. Evaluation of the mandibular canal and the third mandibular molar relationship by CBCT with a deep learning approach. Oral Radiol 2024:10.1007/s11282-024-00793-z. [PMID: 39658743 DOI: 10.1007/s11282-024-00793-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2024] [Accepted: 11/27/2024] [Indexed: 12/12/2024]
Abstract
OBJECTIVE The mandibular canal (MC) houses the inferior alveolar nerve. Extraction of the mandibular third molar (MM3) is a common dental surgery, often complicated by nerve damage. CBCT is the most effective imaging method to assess the relationship between MM3 and MC. With advancements in artificial intelligence, deep learning has shown promising results in dentistry. The aim of this study is to evaluate the MC-MM3 relationship using CBCT and a deep learning technique, as well as to automatically segment the mandibular impacted third molar, mandibular canal, mental and mandibular foramen. METHODS This retrospective study analyzed CBCT data from 300 patients. Segmentation was used for labeling, dividing the data into training (n = 270) and test (n = 30) sets. The nnU-NetV2 architecture was employed to develop an optimal deep learning model. The model's success was validated using the test set, with metrics including accuracy, sensitivity, precision, Dice score, Jaccard index, and AUC. RESULTS For the MM3 annotated on CBCT, the accuracy was 0.99, sensitivity 0.90, precision 0.85, Dice score 0.85, Jaccard index 0.78, AUC value 0.95. In MC evaluation, accuracy was 0.99, sensitivity 0.75, precision 0.78, Dice score 0.76, Jaccard index 0.62, AUC value 0.88. For the evaluation of mental foramen; accuracy 0.99, sensitivity 0.64, precision 0.66, Dice score 0.64, Jaccard index 0.57, AUC value 0.82. In the evaluation of mandibular foramen, accuracy was found to be 0.99, sensitivity 0.79, precision 0.68, Dice score 0.71, and AUC value 0.90. Evaluating the MM3-MC relationship, the model showed an 80% correlation with observer assessments. CONCLUSION The nnU-NetV2 deep learning architecture reliably identifies the MC-MM3 relationship in CBCT images, aiding in diagnosis, surgical planning, and complication prediction.
Collapse
Affiliation(s)
- Suay Yağmur Ünal
- Faculty of Dentistry, Department of Oral and Maxillofacial Radiology, Marmara University, Başıbüyük, Başıbüyük Başıbüyük Yolu Marmara Üniversitesi, Sağlık Yerleşkesi 9/3, 34854, Maltepe, Istanbul, Turkey.
| | - Filiz Namdar Pekiner
- Faculty of Dentistry, Department of Oral and Maxillofacial Radiology, Marmara University, Başıbüyük, Başıbüyük Başıbüyük Yolu Marmara Üniversitesi, Sağlık Yerleşkesi 9/3, 34854, Maltepe, Istanbul, Turkey
| |
Collapse
|
4
|
Chen Y, Du P, Zhang Y, Guo X, Song Y, Wang J, Yang LL, He W. Image-based multi-omics analysis for oral science: Recent progress and perspectives. J Dent 2024; 151:105425. [PMID: 39427959 DOI: 10.1016/j.jdent.2024.105425] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2024] [Revised: 10/01/2024] [Accepted: 10/18/2024] [Indexed: 10/22/2024] Open
Abstract
OBJECTIVES The diagnosis and treatment of oral and dental diseases rely heavily on various types of medical imaging. Deep learning-mediated multi-omics analysis can extract more representative features than those identified through traditional diagnostic methods. This review aims to discuss the applications and recent advances in image-based multi-omics analysis in oral science and to highlight its potential to enhance traditional diagnostic approaches for oral diseases. STUDY SELECTION, DATA, AND SOURCES A systematic search was conducted in the PubMed, Web of Science, and Google Scholar databases, covering all available records. This search thoroughly examined and summarized advances in image-based multi-omics analysis in oral and maxillofacial medicine. CONCLUSIONS This review comprehensively summarizes recent advancements in image-based multi-omics analysis for oral science, including radiomics, pathomics, and photographic-based omics analysis. It also discusses the ongoing challenges and future perspectives that could provide new insights into exploiting the potential of image-based omics analysis in the field of oral science. CLINICAL SIGNIFICANCE This review article presents the state of image-based multi-omics analysis in stomatology, aiming to help oral clinicians recognize the utility of combining omics analyses with imaging during diagnosis and treatment, which can improve diagnostic accuracy, shorten times to diagnosis, save medical resources, and reduce disparity in professional knowledge among clinicians.
Collapse
Affiliation(s)
- Yizhuo Chen
- Department of Stomatology, The First Affiliated Hospital of Zhengzhou University, Zhengzhou 450052, China
| | - Pengxi Du
- Department of Stomatology, The First Affiliated Hospital of Zhengzhou University, Zhengzhou 450052, China
| | - Yinyin Zhang
- Department of Stomatology, The First Affiliated Hospital of Zhengzhou University, Zhengzhou 450052, China
| | - Xin Guo
- Department of Stomatology, The First Affiliated Hospital of Zhengzhou University, Zhengzhou 450052, China
| | - Yujing Song
- Department of Stomatology, The First Affiliated Hospital of Zhengzhou University, Zhengzhou 450052, China
| | - Jianhua Wang
- Department of Stomatology, The First Affiliated Hospital of Zhengzhou University, Zhengzhou 450052, China
| | - Lei-Lei Yang
- Department of Stomatology, The First Affiliated Hospital of Zhengzhou University, Zhengzhou 450052, China.
| | - Wei He
- Department of Stomatology, The First Affiliated Hospital of Zhengzhou University, Zhengzhou 450052, China.
| |
Collapse
|
5
|
Revilla-León M, Gómez-Polo M, Sailer I, Kois JC, Rokhshad R. An overview of artificial intelligence based applications for assisting digital data acquisition and implant planning procedures. J ESTHET RESTOR DENT 2024; 36:1666-1674. [PMID: 38757761 DOI: 10.1111/jerd.13249] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2024] [Revised: 04/29/2024] [Accepted: 04/30/2024] [Indexed: 05/18/2024]
Abstract
OBJECTIVES To provide an overview of the current artificial intelligence (AI) based applications for assisting digital data acquisition and implant planning procedures. OVERVIEW A review of the main AI-based applications integrated into digital data acquisitions technologies (facial scanners (FS), intraoral scanners (IOSs), cone beam computed tomography (CBCT) devices, and jaw trackers) and computer-aided static implant planning programs are provided. CONCLUSIONS The main AI-based application integrated in some FS's programs involves the automatic alignment of facial and intraoral scans for virtual patient integration. The AI-based applications integrated into IOSs programs include scan cleaning, assist scanning, and automatic alignment between the implant scan body with its corresponding CAD object while scanning. The more frequently AI-based applications integrated into the programs of CBCT units involve positioning assistant, noise and artifacts reduction, structures identification and segmentation, airway analysis, and alignment of facial, intraoral, and CBCT scans. Some computer-aided static implant planning programs include patient's digital files, identification, labeling, and segmentation of anatomical structures, mandibular nerve tracing, automatic implant placement, and surgical implant guide design.
Collapse
Affiliation(s)
- Marta Revilla-León
- Department of Restorative Dentistry, School of Dentistry, University of Washington, Seattle, Washington, USA
- Research and Digital Dentistry, Kois Center, Seattle, Washington, USA
- Department of Prosthodontics, School of Dental Medicine, Tufts University, Boston, Massachusetts, USA
| | - Miguel Gómez-Polo
- Department of Conservative Dentistry and Prosthodontics, Complutense University of Madrid, Madrid, Spain
- Advanced in Implant-Prosthodontics, School of Dentistry, Complutense University of Madrid, Madrid, Spain
| | - Irena Sailer
- Fixed Prosthodontics and Biomaterials, University Clinic of Dental Medicine, University of Geneva, Geneva, Switzerland
| | - John C Kois
- Kois Center, Seattle, Washington, USA
- Department of Restorative Dentistry, University of Washington, Seattle, Washington, USA
- Private Practice, Seattle, Washington, USA
| | - Rata Rokhshad
- Topic Group Dental Diagnostics and Digital Dentistry, ITU/WHO Focus Group AI on Health, Berlin, Germany
| |
Collapse
|
6
|
Chen W, Dhawan M, Liu J, Ing D, Mehta K, Tran D, Lawrence D, Ganhewa M, Cirillo N. Mapping the Use of Artificial Intelligence-Based Image Analysis for Clinical Decision-Making in Dentistry: A Scoping Review. Clin Exp Dent Res 2024; 10:e70035. [PMID: 39600121 PMCID: PMC11599430 DOI: 10.1002/cre2.70035] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/19/2024] [Revised: 09/19/2024] [Accepted: 10/20/2024] [Indexed: 11/29/2024] Open
Abstract
OBJECTIVES Artificial intelligence (AI) is an emerging field in dentistry. AI is gradually being integrated into dentistry to improve clinical dental practice. The aims of this scoping review were to investigate the application of AI in image analysis for decision-making in clinical dentistry and identify trends and research gaps in the current literature. MATERIAL AND METHODS This review followed the guidelines provided by the Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews (PRISMA-ScR). An electronic literature search was performed through PubMed and Scopus. After removing duplicates, a preliminary screening based on titles and abstracts was performed. A full-text review and analysis were performed according to predefined inclusion criteria, and data were extracted from eligible articles. RESULTS Of the 1334 articles returned, 276 met the inclusion criteria (consisting of 601,122 images in total) and were included in the qualitative synthesis. Most of the included studies utilized convolutional neural networks (CNNs) on dental radiographs such as orthopantomograms (OPGs) and intraoral radiographs (bitewings and periapicals). AI was applied across all fields of dentistry - particularly oral medicine, oral surgery, and orthodontics - for direct clinical inference and segmentation. AI-based image analysis was use in several components of the clinical decision-making process, including diagnosis, detection or classification, prediction, and management. CONCLUSIONS A variety of machine learning and deep learning techniques are being used for dental image analysis to assist clinicians in making accurate diagnoses and choosing appropriate interventions in a timely manner.
Collapse
Affiliation(s)
- Wei Chen
- Melbourne Dental SchoolThe University of MelbourneCarltonVictoriaAustralia
| | - Monisha Dhawan
- Melbourne Dental SchoolThe University of MelbourneCarltonVictoriaAustralia
| | - Jonathan Liu
- Melbourne Dental SchoolThe University of MelbourneCarltonVictoriaAustralia
| | - Damie Ing
- Melbourne Dental SchoolThe University of MelbourneCarltonVictoriaAustralia
| | - Kruti Mehta
- Melbourne Dental SchoolThe University of MelbourneCarltonVictoriaAustralia
| | - Daniel Tran
- Melbourne Dental SchoolThe University of MelbourneCarltonVictoriaAustralia
| | | | - Max Ganhewa
- CoTreatAI, CoTreat Pty Ltd.MelbourneVictoriaAustralia
| | - Nicola Cirillo
- Melbourne Dental SchoolThe University of MelbourneCarltonVictoriaAustralia
- CoTreatAI, CoTreat Pty Ltd.MelbourneVictoriaAustralia
| |
Collapse
|
7
|
Huang C, Wang Y, Wang Y, Zhao Z. Automatic detection and proximity quantification of inferior alveolar nerve and mandibular third molar on cone-beam computed tomography. Clin Oral Investig 2024; 28:648. [PMID: 39567447 DOI: 10.1007/s00784-024-05967-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2024] [Accepted: 09/24/2024] [Indexed: 11/22/2024]
Abstract
OBJECTIVES During mandibular third molar (MTM) extraction surgery, preoperative analysis to quantify the proximity of the MTM to the surrounding inferior alveolar nerve (IAN) is essential to minimize the risk of IAN injury. This study aims to propose an automated tool to quantitatively measure the proximity of IAN and MTM in cone-beam computed tomography (CBCT) images. MATERIALS AND METHODS Using the dataset including 302 CBCT scans with 546 MTMs, a deep-learning-based network was developed to support the automatic detection of the IAN, MTM, and intersection region IR. To ensure accurate proximity detection, a distance detection algorithm and a volume measurement algorithm were also developed. RESULTS The deep learning-based model showed encouraging segmentation accuracy of the target structures (Dice similarity coefficient: 0.9531 ± 0.0145, IAN; 0.9832 ± 0.0055, MTM; 0.8336 ± 0.0746, IR). In addition, with the application of the developed algorithms, the distance between the IAN and MTM and the volume of the IR could be equivalently detected (90% confidence interval (CI): - 0.0345-0.0014 mm, distance; - 0.0155-0.0759 mm3, volume). The total time for the IAN, MTM, and IR segmentation was 2.96 ± 0.11 s, while the accurate manual segmentation required 39.01 ± 5.89 min. CONCLUSIONS This study presented a novel, fast, and accurate model for the detection and proximity quantification of the IAN and MTM on CBCT. CLINICAL RELEVANCE This model illustrates that a deep learning network may assist surgeons in evaluating the risk of MTM extraction surgery by detecting the proximity of the IAN and MTM at a quantitative level that was previously unparalleled.
Collapse
Affiliation(s)
- Chao Huang
- State Key Laboratory of Oral Diseases &, National Clinical Research Center for Oral Diseases, West China Hospital of Stomatology, Sichuan University, Chengdu, China
- Department of Orthodontics, West China Hospital of Stomatology, Sichuan University, Chengdu, China
| | - Yigan Wang
- State Key Laboratory of Oral Diseases &, National Clinical Research Center for Oral Diseases, West China Hospital of Stomatology, Sichuan University, Chengdu, China
- Department of Orthodontics, West China Hospital of Stomatology, Sichuan University, Chengdu, China
| | - Yifan Wang
- State Key Laboratory of Oral Diseases &, National Clinical Research Center for Oral Diseases, West China Hospital of Stomatology, Sichuan University, Chengdu, China
- Department of Orthodontics, West China Hospital of Stomatology, Sichuan University, Chengdu, China
| | - Zhihe Zhao
- State Key Laboratory of Oral Diseases &, National Clinical Research Center for Oral Diseases, West China Hospital of Stomatology, Sichuan University, Chengdu, China.
- Department of Orthodontics, West China Hospital of Stomatology, Sichuan University, Chengdu, China.
| |
Collapse
|
8
|
Soltani P, Sohrabniya F, Mohammad-Rahimi H, Mehdizadeh M, Mohammadreza Mousavi S, Moaddabi A, Mohammadmahdi Mousavi S, Spagnuolo G, Yavari A, Schwendicke F. A two-stage deep-learning model for determination of the contact of mandibular third molars with the mandibular canal on panoramic radiographs. BMC Oral Health 2024; 24:1373. [PMID: 39538183 PMCID: PMC11562527 DOI: 10.1186/s12903-024-04850-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2024] [Accepted: 09/02/2024] [Indexed: 11/16/2024] Open
Abstract
OBJECTIVES This study aimed to assess the accuracy of a two-stage deep learning (DL) model for (1) detecting mandibular third molars (MTMs) and the mandibular canal (MC), and (2) classifying the anatomical relationship between these structures (contact/no contact) on panoramic radiographs. METHOD MTMs and MCs were labeled on panoramic radiographs by a calibrated examiner using bounding boxes. Each bounding box contained MTM and MC on one side. The relationship of MTMs with the MC was assessed on CBCT scans by two independent examiners without the knowledge of the condition of MTM and MC on the corresponding panoramic image, and dichotomized as contact/no contact. Data were split into training, validation, and testing sets with a ratio of 80:10:10. Faster R-CNN was used for detecting MTMs and MCs and ResNeXt for classifying their relationship. AP50 and AP75 were used as outcomes for detecting MTMs and MCs, and accuracy, precision, recall, F1-score, and the area-under-the-receiver-operating-characteristics curve (AUROC) were used to assess classification performance. The training and validation of the models were conducted using the Python programming language with the PyTorch framework. RESULTS Three hundred eighty-seven panoramic radiographs were evaluated. MTMs were present bilaterally on 232 and unilaterally on 155 radiographs. In total, 619 images were collected which included MTMs and MCs. AP50 and AP75 indicating accuracy for detecting MTMs and MCs were 0.99 and 0.90 respectively. Classification accuracy, recall, specificity, F1-score, precision, and AUROC values were 0.85, 0.85, 0.93, 0.84, 0.86, and 0.91, respectively. CONCLUSION DL can detect MTMs and MCs and accurately assess their anatomical relationship on panoramic radiographs.
Collapse
Affiliation(s)
- Parisa Soltani
- Department of Oral and Maxillofacial Radiology, Dental Implants Research Center, Dental Research Institute, School of Dentistry, Isfahan University of Medical Sciences, Isfahan, Iran
- Department of Neurosciences, Reproductive and Odontostomatological Sciences, University of Naples "Federico II", Naples, Italy
| | - Fatemeh Sohrabniya
- Topic Group Dental Diagnostics and Digital Dentistry, ITU/WHO Focus Group AI on Health, Berlin, Federal Republic of Germany
| | - Hossein Mohammad-Rahimi
- Topic Group Dental Diagnostics and Digital Dentistry, ITU/WHO Focus Group AI on Health, Berlin, Federal Republic of Germany
| | - Mojdeh Mehdizadeh
- Department of Oral and Maxillofacial Radiology, Dental Implants Research Center, Dental Research Institute, School of Dentistry, Isfahan University of Medical Sciences, Isfahan, Iran
| | | | - Amirhossein Moaddabi
- Department of Oral and Maxillofacial Surgery, Dental Research Center, Mazandaran University of Medical Sciences, Sari, Iran.
| | | | - Gianrico Spagnuolo
- Department of Neurosciences, Reproductive and Odontostomatological Sciences, University of Naples "Federico II", Naples, Italy
| | - Amirmohammad Yavari
- Students Research Committee, School of Dentistry, Isfahan University of Medical Sciences, Isfahan, Iran
| | - Falk Schwendicke
- Clinic for Conservative Dentistry and Periodontology, University Hospital of the Ludwig-Maximilians- University Munich, Munich, Germany
| |
Collapse
|
9
|
Barnes NA, Dkhar W, S S, Chhaparwal Y, Mayya V, H RC. Automated classification of mandibular canal in relation to third molar using CBCT images. F1000Res 2024; 13:995. [PMID: 39926002 PMCID: PMC11803391 DOI: 10.12688/f1000research.154985.1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Accepted: 08/28/2024] [Indexed: 02/11/2025] Open
Abstract
Background Dental radiology has significantly benefited from cone-beam computed tomography (CBCT) because of its compact size and low radiation exposure. Canal tracking is an important application of CBCT for determining the relationship between the inferior alveolar nerve and third molar. Usually, canal tacking is performed manually, which takes a lot of time. This study aimed to develop an artificial intelligence (AI) model to automate classification of the mandibular canal in relation to the third molar. Methods This retrospective study was conducted using 434 CBCT images. 3D slicer software was used to annotate and classify the data into lingual, buccal, and inferior categories. Two convolution neural network models, AlexNet and ResNet50, were developed to classify this relationship. The study included 262 images for training and 172 images for testing, with the model performance evaluated by sensitivity, precision, and F1 score. Results The performance of the two models was evaluated using a 3 × 3 confusion matrix, with the data categorized into 3 clases: lingual, buccal, and inferior. The mandibular canal and third molar have a close anatomical relationship, highlighting the need for precise imaging in dental and surgical settings. To accurately classify the mandibular canal in relation to the third molar, both AlexNet and ResNet50 demonstrated high accuracy, with F1 scores ranging from 0.64 to 0.92 for different classes, with accuracy of 81% and 83%, respectively, for accurately classifying the mandibular canal in relation to the third molar. Conclusion The present study developed and evaluated AI models to accurately classify and establish the relationship between the mandibular canal and third molars using CBCT images with a higher accuracy rate.
Collapse
Affiliation(s)
- Neil Abraham Barnes
- Department of Medical Imaging Technology, Manipal College of Health Professions, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - Winniecia Dkhar
- Department of Medical Imaging Technology, Manipal College of Health Professions, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - Sharath S
- Department of Medical Imaging Technology, Manipal College of Health Professions, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - Yogesh Chhaparwal
- Department of Oral Medicine and Radiology, Manipal College of Dental Sciences, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - Veena Mayya
- Departments of Information & Communication Technology, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - Roopitha C H
- Departments of Information & Communication Technology, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| |
Collapse
|
10
|
Pirayesh Z, Mohammad-Rahimi H, Motamedian SR, Amini Afshar S, Abbasi R, Rohban MH, Mahdian M, Ghazizadeh Ahsaie M, Iranparvar Alamdari M. A hierarchical deep learning approach for diagnosing impacted canine-induced root resorption via cone-beam computed tomography. BMC Oral Health 2024; 24:982. [PMID: 39180070 PMCID: PMC11344340 DOI: 10.1186/s12903-024-04718-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2024] [Accepted: 08/08/2024] [Indexed: 08/26/2024] Open
Abstract
OBJECTIVES Canine-induced root resorption (CIRR) is caused by impacted canines and CBCT images have shown to be more accurate in diagnosing CIRR than panoramic and periapical radiographs with the reported AUCs being 0.95, 0.49, and 0.57, respectively. The aim of this study was to use deep learning to automatically evaluate the diagnosis of CIRR in maxillary incisors using CBCT images. METHODS A total of 50 cone beam computed tomography (CBCT) images and 176 incisors were selected for the present study. The maxillary incisors were manually segmented and labeled from the CBCT images by two independent radiologists as either healthy or affected by root resorption induced by the impacted canines. We used five different strategies for training the model: (A) classification using 3D ResNet50 (Baseline), (B) classification of the segmented masks using the outcome of a 3D U-Net pretrained on the 3D MNIST, (C) training a 3D U-Net for the segmentation task and use its outputs for classification, (D) pretraining a 3D U-Net for the segmentation and transfer of the model, and (E) pretraining a 3D U-Net for the segmentation and fine-tuning the model with only the model encoder. The segmentation models were evaluated using the mean intersection over union (mIoU) and Dice coefficient (DSC). The classification models were evaluated in terms of classification accuracy, precision, recall, and F1 score. RESULTS The segmentation model achieved a mean intersection over union (mIoU) of 0.641 and a DSC of 0.901, indicating good performance in segmenting the tooth structures from the CBCT images. For the main classification task of detecting CIRR, Model C (classification of the segmented masks using 3D ResNet) and Model E (pretraining on segmentation followed by fine-tuning for classification) performed the best, both achieving 82% classification accuracy and 0.62 F1-scores on the test set. These results demonstrate the effectiveness of the proposed hierarchical, data-efficient deep learning approaches in improving the accuracy of automated CIRR diagnosis from limited CBCT data compared to the 3D ResNet baseline model. CONCLUSION The proposed approaches are effective at improving the accuracy of classification tasks and are helpful when the diagnosis is based on the volume and boundaries of an object. While the study demonstrated promising results, future studies with larger sample size are required to validate the effectiveness of the proposed method in enhancing the medical image classification tasks.
Collapse
Affiliation(s)
- Zeynab Pirayesh
- Topic Group Dental Diagnostics and Digital Dentistry, ITU/WHO Focus Group AI on Health, Berlin, Germany
- Department of Orthodontics and Dentofacial Orthopedics, School of Dentistry, Zanjan University of Medical Sciences, Zanjan, Iran
| | - Hossein Mohammad-Rahimi
- Topic Group Dental Diagnostics and Digital Dentistry, ITU/WHO Focus Group AI on Health, Berlin, Germany
- Department of Computer Engineering, Sharif University of Technology, Tehran, Iran
| | - Saeed Reza Motamedian
- Topic Group Dental Diagnostics and Digital Dentistry, ITU/WHO Focus Group AI on Health, Berlin, Germany
- Dentofacial Deformities Research Center, Research Institute of Dental Sciences, Shahid Beheshti University of Medical Sciences, Tehran, Iran
- Department of Orthodontics, School of Dentistry, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Sepehr Amini Afshar
- Department of Computer Engineering, Sharif University of Technology, Tehran, Iran
| | - Reza Abbasi
- Department of Computer Engineering, Sharif University of Technology, Tehran, Iran
| | | | - Mina Mahdian
- Division of Diagnostic Imaging, Department of Prosthodontics and Digital Technology, Stony Brook University School of Dental Medicine, Stony Brook, NY, USA
| | - Mitra Ghazizadeh Ahsaie
- Department of Electrical and Computer Engineering, California State University, Chico, 95929, USA
| | - Mina Iranparvar Alamdari
- Department of Oral and Maxillofacial Radiology, School of Dentistry, Shahid Beheshti University of Medical Sciences, Tehran, Iran.
| |
Collapse
|
11
|
Dong F, Yan J, Zhang X, Zhang Y, Liu D, Pan X, Xue L, Liu Y. Artificial intelligence-based predictive model for guidance on treatment strategy selection in oral and maxillofacial surgery. Heliyon 2024; 10:e35742. [PMID: 39170321 PMCID: PMC11336844 DOI: 10.1016/j.heliyon.2024.e35742] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/05/2024] [Revised: 07/27/2024] [Accepted: 08/02/2024] [Indexed: 08/23/2024] Open
Abstract
Application of deep learning (DL) and machine learning (ML) is rapidly increasing in the medical field. DL is gaining significance for medical image analysis, particularly, in oral and maxillofacial surgeries. Owing to the ability to accurately identify and categorize both diseased and normal soft- and hard-tissue structures, DL has high application potential in the diagnosis and treatment of tumors and in orthognathic surgeries. Moreover, DL and ML can be used to develop prediction models that can aid surgeons to assess prognosis by analyzing the patient's medical history, imaging data, and surgical records, develop more effective treatment strategies, select appropriate surgical modalities, and evaluate the risk of postoperative complications. Such prediction models can play a crucial role in the selection of treatment strategies for oral and maxillofacial surgeries. Their practical application can improve the utilization of medical staff, increase the treatment accuracy and efficiency, reduce surgical risks, and provide an enhanced treatment experience to patients. However, DL and ML face limitations, such as data drift, unstable model results, and vulnerable social trust. With the advancement of social concepts and technologies, the use of these models in oral and maxillofacial surgery is anticipated to become more comprehensive and extensive.
Collapse
Affiliation(s)
- Fanqiao Dong
- School of Stomatology, China Medical University, Shenyang, China
| | - Jingjing Yan
- Hospital of Stomatology, China Medical University, Shenyang, China
| | - Xiyue Zhang
- School of Stomatology, China Medical University, Shenyang, China
| | - Yikun Zhang
- School of Stomatology, China Medical University, Shenyang, China
| | - Di Liu
- School of Stomatology, China Medical University, Shenyang, China
| | - Xiyun Pan
- School of Stomatology, China Medical University, Shenyang, China
| | - Lei Xue
- School of Stomatology, China Medical University, Shenyang, China
- Hospital of Stomatology, China Medical University, Shenyang, China
| | - Yu Liu
- First Affiliated Hospital of Jinzhou Medical University, Jinzhou, China
| |
Collapse
|
12
|
Ni FD, Xu ZN, Liu MQ, Zhang MJ, Li S, Bai HL, Ding P, Fu KY. Towards clinically applicable automated mandibular canal segmentation on CBCT. J Dent 2024; 144:104931. [PMID: 38458378 DOI: 10.1016/j.jdent.2024.104931] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/20/2023] [Revised: 03/04/2024] [Accepted: 03/05/2024] [Indexed: 03/10/2024] Open
Abstract
OBJECTIVES To develop a deep learning-based system for precise, robust, and fully automated segmentation of the mandibular canal on cone beam computed tomography (CBCT) images. METHODS The system was developed on 536 CBCT scans (training set: 376, validation set: 80, testing set: 80) from one center and validated on an external dataset of 89 CBCT scans from 3 centers. Each scan was annotated using a multi-stage annotation method and refined by oral and maxillofacial radiologists. We proposed a three-step strategy for the mandibular canal segmentation: extraction of the region of interest based on 2D U-Net, global segmentation of the mandibular canal, and segmentation refinement based on 3D U-Net. RESULTS The system consistently achieved accurate mandibular canal segmentation in the internal set (Dice similarity coefficient [DSC], 0.952; intersection over union [IoU], 0.912; average symmetric surface distance [ASSD], 0.046 mm; 95% Hausdorff distance [HD95], 0.325 mm) and the external set (DSC, 0.960; IoU, 0.924; ASSD, 0.040 mm; HD95, 0.288 mm). CONCLUSIONS These results demonstrated the potential clinical application of this AI system in facilitating clinical workflows related to mandibular canal localization. CLINICAL SIGNIFICANCE Accurate delineation of the mandibular canal on CBCT images is critical for implant placement, mandibular third molar extraction, and orthognathic surgery. This AI system enables accurate segmentation across different models, which could contribute to more efficient and precise dental automation systems.
Collapse
Affiliation(s)
- Fang-Duan Ni
- Department of Oral & Maxillofacial Radiology, Peking University School & Hospital of Stomatology, Beijing 100081, China; National Center for Stomatology & National Clinical Research Center for Oral Diseases, Beijing 100081, China; National Engineering Research Center of Oral Biomaterials and Digital Medical Devices, Beijing 100081, China; Beijing Key Laboratory of Digital Stomatology, Beijing 100081, China
| | | | - Mu-Qing Liu
- Department of Oral & Maxillofacial Radiology, Peking University School & Hospital of Stomatology, Beijing 100081, China; National Center for Stomatology & National Clinical Research Center for Oral Diseases, Beijing 100081, China; National Engineering Research Center of Oral Biomaterials and Digital Medical Devices, Beijing 100081, China; Beijing Key Laboratory of Digital Stomatology, Beijing 100081, China.
| | - Min-Juan Zhang
- Second Dental Center, Peking University Hospital of Stomatology, Beijing 100101, China
| | - Shu Li
- Department of Stomatology, Beijing Hospital, Beijing 100005, China
| | | | | | - Kai-Yuan Fu
- Department of Oral & Maxillofacial Radiology, Peking University School & Hospital of Stomatology, Beijing 100081, China; National Center for Stomatology & National Clinical Research Center for Oral Diseases, Beijing 100081, China; National Engineering Research Center of Oral Biomaterials and Digital Medical Devices, Beijing 100081, China; Beijing Key Laboratory of Digital Stomatology, Beijing 100081, China.
| |
Collapse
|
13
|
Tyndall DA, Price JB, Gaalaas L, Spin-Neto R. Surveying the landscape of diagnostic imaging in dentistry's future: Four emerging technologies with promise. J Am Dent Assoc 2024; 155:364-378. [PMID: 38520421 DOI: 10.1016/j.adaj.2024.01.005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/24/2023] [Revised: 01/04/2024] [Accepted: 01/07/2024] [Indexed: 03/25/2024]
Abstract
BACKGROUND Advances in digital radiography for both intraoral and panoramic imaging and cone-beam computed tomography have led the way to an increase in diagnostic capabilities for the dental care profession. In this article, the authors provide information on 4 emerging technologies with promise. TYPES OF STUDIES REVIEWED The authors feature the following: artificial intelligence in the form of deep learning using convolutional neural networks, dental magnetic resonance imaging, stationary intraoral tomosynthesis, and second-generation cone-beam computed tomography sources based on carbon nanotube technology and multispectral imaging. The authors review and summarize articles featuring these technologies. RESULTS The history and background of these emerging technologies are previewed along with their development and potential impact on the practice of dental diagnostic imaging. The authors conclude that these emerging technologies have the potential to have a substantial influence on the practice of dentistry as these systems mature. The degree of influence most likely will vary, with artificial intelligence being the most influential of the 4. CONCLUSIONS AND PRACTICAL IMPLICATIONS The readers are informed about these emerging technologies and the potential effects on their practice going forward, giving them information on which to base decisions on adopting 1 or more of these technologies. The 4 technologies reviewed in this article have the potential to improve imaging diagnostics in dentistry thereby leading to better patient care and heightened professional satisfaction.
Collapse
|
14
|
Wei L, Wu S, Huang Z, Chen Y, Zheng H, Wang L. Autologous Transplantation Tooth Guide Design Based on Deep Learning. J Oral Maxillofac Surg 2024; 82:314-324. [PMID: 37832596 DOI: 10.1016/j.joms.2023.09.014] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2023] [Revised: 09/14/2023] [Accepted: 09/14/2023] [Indexed: 10/15/2023]
Abstract
BACKGROUND Autologous tooth transplantation requires precise surgical guide design, involving manual tracing of donor tooth contours based on patient cone-beam computed tomography (CBCT) scans. While manual corrections are time-consuming and prone to human errors, deep learning-based approaches show promise in reducing labor and time costs while minimizing errors. However, the application of deep learning techniques in this particular field is yet to be investigated. PURPOSE We aimed to assess the feasibility of replacing the traditional design pipeline with a deep learning-enabled autologous tooth transplantation guide design pipeline. STUDY DESIGN, SETTING, SAMPLE This retrospective cross-sectional study used 79 CBCT images collected at the Guangzhou Medical University Hospital between October 2022 and March 2023. Following preprocessing, a total of 5,070 region of interest images were extracted from 79 CBCT images. PREDICTOR VARIABLE Autologous tooth transplantation guide design pipelines, either based on traditional manual design or deep learning-based design. MAIN OUTCOME VARIABLE The main outcome variable was the error between the reconstructed model and the gold standard benchmark. We used the third molar extracted clinically as the gold standard and leveraged it as the benchmark for evaluating our reconstructed models from different design pipelines. Both trueness and accuracy were used to evaluate this error. Trueness was assessed using the root mean square (RMS), and accuracy was measured using the standard deviation. The secondary outcome variable was the pipeline efficiency, assessed based on the time cost. Time cost refers to the amount of time required to acquire the third molar model using the pipeline. ANALYSES Data were analyzed using the Kruskal-Wallis test. Statistical significance was set at P < .05. RESULTS In the surface matching comparison for different reconstructed models, the deep learning group achieved the lowest RMS value (0.335 ± 0.066 mm). There were no significant differences in RMS values between manual design by a senior doctor and deep learning-based design (P = .688), and the standard deviation values did not differ among the 3 groups (P = .103). The deep learning-based design pipeline (0.017 ± 0.001 minutes) provided a faster assessment compared to the manual design pipeline by both senior (19.676 ± 2.386 minutes) and junior doctors (30.613 ± 6.571 minutes) (P < .001). CONCLUSIONS AND RELEVANCE The deep learning-based automatic pipeline exhibited similar performance in surgical guide design for autogenous tooth transplantation compared to manual design by senior doctors, and it minimized time costs.
Collapse
Affiliation(s)
- Lifen Wei
- Department of Dental Implantation, Affiliated Stomatology Hospital of Guangzhou Medical University, Guangdong Engineering Research Center of Oral Restoration and Reconstruction, Guangzhou Key Laboratory of Basic and Applied Research of Oral Regenerative Medicine, Guangzhou, Guangdong, China
| | - Shuyang Wu
- Department of Pathology, Collaborative Innovation Center for Cancer Medicine, State Key Laboratory of Oncology in South China, Sun Yat-sen University Cancer Center, Guangzhou, Guangdong, China
| | - Zelun Huang
- Department of Dental Implantation, Affiliated Stomatology Hospital of Guangzhou Medical University, Guangdong Engineering Research Center of Oral Restoration and Reconstruction, Guangzhou Key Laboratory of Basic and Applied Research of Oral Regenerative Medicine, Guangzhou, Guangdong, China
| | - Yaxin Chen
- Department of Dental Implantation, Affiliated Stomatology Hospital of Guangzhou Medical University, Guangdong Engineering Research Center of Oral Restoration and Reconstruction, Guangzhou Key Laboratory of Basic and Applied Research of Oral Regenerative Medicine, Guangzhou, Guangdong, China
| | - Haoran Zheng
- Department of Chemical & Materials Engineering, University of Auckland, Auckland, New Zealand
| | - Liping Wang
- Department of Dental Implantation, Affiliated Stomatology Hospital of Guangzhou Medical University, Guangdong Engineering Research Center of Oral Restoration and Reconstruction, Guangzhou Key Laboratory of Basic and Applied Research of Oral Regenerative Medicine, Guangzhou, Guangdong, China.
| |
Collapse
|
15
|
Faadiya AN, Widyaningrum R, Arindra PK, Diba SF. The diagnostic performance of impacted third molars in the mandible: A review of deep learning on panoramic radiographs. Saudi Dent J 2024; 36:404-412. [PMID: 38525176 PMCID: PMC10960107 DOI: 10.1016/j.sdentj.2023.11.025] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2023] [Revised: 11/21/2023] [Accepted: 11/23/2023] [Indexed: 03/26/2024] Open
Abstract
Background Mandibular third molar is prone to impaction, resulting in its inability to erupt into the oral cavity. The radiographic examination is required to support the odontectomy of impacted teeth. The use of computer-aided diagnosis based on deep learning is emerging in the field of medical and dentistry with the advancement of artificial intelligence (AI) technology. This review describes the performance and prospects of deep learning for the detection, classification, and evaluation of third molar-mandibular canal relationships on panoramic radiographs. Methods This work was conducted using three databases: PubMed, Google Scholar, and Science Direct. Following the literature selection, 49 articles were reviewed, with the 12 main articles discussed in this review. Results Several models of deep learning are currently used for segmentation and classification of third molar impaction with or without the combination of other techniques. Deep learning has demonstrated significant diagnostic performance in identifying mandibular impacted third molars (ITM) on panoramic radiographs, with an accuracy range of 78.91% to 90.23%. Meanwhile, the accuracy of deep learning in determining the relationship between ITM and the mandibular canal (MC) ranges from 72.32% to 99%. Conclusion Deep learning-based AI with high performance for the detection, classification, and evaluation of the relationship of ITM to the MC using panoramic radiographs has been developed over the past decade. However, deep learning must be improved using large datasets, and the evaluation of diagnostic performance for deep learning models should be aligned with medical diagnostic test protocols. Future studies involving collaboration among oral radiologists, clinicians, and computer scientists are required to identify appropriate AI development models that are accurate, efficient, and applicable to clinical services.
Collapse
Affiliation(s)
- Amalia Nur Faadiya
- Dental Medicine Study Program, Faculty of Dentistry, Universitas Gadjah Mada, Yogyakarta, Indonesia
| | - Rini Widyaningrum
- Department of Dentomaxillofacial Radiology, Faculty of Dentistry, Universitas Gadjah Mada, Yogyakarta, Indonesia
| | - Pingky Krisna Arindra
- Department of Oral and Maxillofacial Surgery, Faculty of Dentistry, Universitas Gadjah Mada, Yogyakarta, Indonesia
| | - Silviana Farrah Diba
- Department of Dentomaxillofacial Radiology, Faculty of Dentistry, Universitas Gadjah Mada, Yogyakarta, Indonesia
| |
Collapse
|
16
|
Gong Z, Feng W, Su X, Choi C. System for automatically assessing the likelihood of inferior alveolar nerve injury. Comput Biol Med 2024; 169:107923. [PMID: 38199211 DOI: 10.1016/j.compbiomed.2024.107923] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/20/2023] [Revised: 12/20/2023] [Accepted: 01/01/2024] [Indexed: 01/12/2024]
Abstract
Inferior alveolar nerve (IAN) injury is a severe complication associated with mandibular third molar (MM3) extraction. Consequently, the likelihood of IAN injury must be assessed before performing such an extraction. However, existing deep learning methods for classifying the likelihood of IAN injury that rely on mask images often suffer from limited accuracy and lack of interpretability. In this paper, we propose an automated system based on panoramic radiographs, featuring a novel segmentation model SS-TransUnet and classification algorithm CD-IAN injury class. Our objective was to enhance the precision of segmentation of MM3 and mandibular canal (MC) and classification accuracy of the likelihood of IAN injury, ultimately reducing the occurrence of IAN injuries and providing a certain degree of interpretable foundation for diagnosis. The proposed segmentation model demonstrated a 0.9 % and 2.6 % enhancement in dice coefficient for MM3 and MC, accompanied by a reduction in 95 % Hausdorff distance, reaching 1.619 and 1.886, respectively. Additionally, our classification algorithm achieved an accuracy of 0.846, surpassing deep learning-based models by 3.8 %, confirming the effectiveness of our system.
Collapse
Affiliation(s)
- Ziyang Gong
- Department of Computer Engineering, Gachon University, Seongnam-si, 13120, Republic of Korea
| | - Weikang Feng
- College of Information Science and Engineering, Hohai University, Changzhou, 213000, China
| | - Xin Su
- College of Information Science and Engineering, Hohai University, Changzhou, 213000, China
| | - Chang Choi
- Department of Computer Engineering, Gachon University, Seongnam-si, 13120, Republic of Korea.
| |
Collapse
|
17
|
Li R, Zhu C, Chu F, Yu Q, Fan D, Ouyang N, Jin Y, Guo W, Xia L, Feng Q, Fang B. Deep learning for virtual orthodontic bracket removal: tool establishment and application. Clin Oral Investig 2024; 28:121. [PMID: 38280038 DOI: 10.1007/s00784-023-05440-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/04/2022] [Accepted: 11/15/2023] [Indexed: 01/29/2024]
Abstract
OBJECTIVE We aimed to develop a tool for virtual orthodontic bracket removal based on deep learning algorithms for feature extraction from bonded teeth and to demonstrate its application in a bracket position assessment scenario. MATERIALS AND METHODS Our segmentation network for virtual bracket removal was trained using dataset A, containing 978 bonded teeth, 20 original teeth, and 20 brackets generated by scanners. The accuracy and segmentation time of the network were tested by dataset B, which included an additional 118 bonded teeth without knowing the original tooth morphology. This tool was then applied for bracket position assessment. The clinical crown center, bracket center, and orientations of separated teeth and brackets were extracted for analyzing the linear distribution and angular deviation of bonded brackets. RESULTS This tool performed virtual bracket removal in 2.9 ms per tooth with accuracies of 98.93% and 97.42% (P < 0.01) in datasets A and B, respectively. The tooth surface and bracket characteristics were extracted and used to evaluate the results of manually bonded brackets by 49 orthodontists. Personal preferences for bracket angulation and bracket distribution were displayed graphically and tabularly. CONCLUSIONS The tool's efficiency and precision are satisfactory, and it can be operated without original tooth data. It can be used to display the bonding deviation in the bracket position assessment scenario. CLINICAL SIGNIFICANCE With the aid of this tool, unnecessary bracket removal can be avoided when evaluating bracket positions and modifying treatment plans. It has the potential to produce retainers and orthodontic devices prior to tooth debonding.
Collapse
Affiliation(s)
- Ruomei Li
- Department of Orthodontics, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai Jiao Tong University, 500 Quxi Road, Shanghai, 200011, China
| | - Cheng Zhu
- Department of Orthodontics, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai Jiao Tong University, 500 Quxi Road, Shanghai, 200011, China
| | - Fengting Chu
- Department of Orthodontics, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai Jiao Tong University, 500 Quxi Road, Shanghai, 200011, China
| | - Quan Yu
- Department of Orthodontics, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai Jiao Tong University, 500 Quxi Road, Shanghai, 200011, China
| | - Di Fan
- School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China
| | - Ningjuan Ouyang
- Department of Orthodontics, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai Jiao Tong University, 500 Quxi Road, Shanghai, 200011, China
| | - Yu Jin
- Department of Orthodontics, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai Jiao Tong University, 500 Quxi Road, Shanghai, 200011, China
| | - Weiming Guo
- Department of Orthodontics, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai Jiao Tong University, 500 Quxi Road, Shanghai, 200011, China
| | - Lunguo Xia
- Department of Orthodontics, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai Jiao Tong University, 500 Quxi Road, Shanghai, 200011, China.
| | - Qiping Feng
- Department of Orthodontics, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai Jiao Tong University, 500 Quxi Road, Shanghai, 200011, China.
| | - Bing Fang
- Department of Orthodontics, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai Jiao Tong University, 500 Quxi Road, Shanghai, 200011, China.
| |
Collapse
|
18
|
de Lima DM, Estrela CRDA, Bernardes CMR, Estrela LRDA, Bueno MR, Estrela C. Spatial Position and Anatomical Characteristics Associated with Impacted Third Molars Using a Map-Reading Strategy on Cone-Beam Computed Tomography Scans: A Retrospective Analysis. Diagnostics (Basel) 2024; 14:260. [PMID: 38337776 PMCID: PMC10855352 DOI: 10.3390/diagnostics14030260] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/21/2023] [Revised: 12/22/2023] [Accepted: 12/28/2023] [Indexed: 02/12/2024] Open
Abstract
(1) Background: This study assessed the spatial position and anatomical features associated with impacted third molars through a map-reading strategy employing cone-beam computed tomography (CBCT). (2) Methods: The positioning of impacted third molars on CBCT was assessed using Winter's and Pell and Gregory's classifications. External root resorption in mandibular second molars was categorized according to Herman's classification. Additionally, the relationship between the mandibular third molar root apex and the mandibular canal was examined. Comparative statistical analysis was conducted using Fisher's exact test, with a significance level considered as 5%. (3) Results: The results indicated that, based on Winter's classification, 48.06 % of impacted teeth were positioned mesioangularly. Employing Pell and Gregory's classification, 43.22% of the impacted molars fell into positions B and C, with 54.2% classified as Class II. A notable 69.7% of teeth exhibited no contact between the root apex and the mandibular canal, and external root resorption in the distal aspect of the second molar was absent in 88.7% of cases. (4) Conclusions: Utilizing the map-reading strategy with CBCT scans to assess the anatomical positions and characteristics of impacted third molars enhances professional confidence and sets a standard for quality and safety in the surgical procedure for patients.
Collapse
Affiliation(s)
- Djalma Maciel de Lima
- Department of Oral Biology, School of Dentistry, Evangelical University of Goiás, Anápolis 75083-515, Brazil; (D.M.d.L.); (C.M.R.B.); (L.R.d.A.E.)
| | - Cyntia Rodrigues de Araújo Estrela
- Department of Oral Biology, School of Dentistry, Evangelical University of Goiás, Anápolis 75083-515, Brazil; (D.M.d.L.); (C.M.R.B.); (L.R.d.A.E.)
| | | | - Lucas Rodrigues de Araújo Estrela
- Department of Oral Biology, School of Dentistry, Evangelical University of Goiás, Anápolis 75083-515, Brazil; (D.M.d.L.); (C.M.R.B.); (L.R.d.A.E.)
| | - Mike Reis Bueno
- Center for Radiology and Orofacial Images, Diagnostic Imaging Center, Cuiabá 78043-272, Brazil;
| | - Carlos Estrela
- Department of Endodontics, School of Dentistry, Federal University of Goiás, Goiânia 74605-020, Brazil;
| |
Collapse
|
19
|
Jeon KJ, Choi H, Lee C, Han SS. Automatic diagnosis of true proximity between the mandibular canal and the third molar on panoramic radiographs using deep learning. Sci Rep 2023; 13:22022. [PMID: 38086921 PMCID: PMC10716248 DOI: 10.1038/s41598-023-49512-4] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2023] [Accepted: 12/08/2023] [Indexed: 12/18/2023] Open
Abstract
Evaluating the mandibular canal proximity is crucial for planning mandibular third molar extractions. Panoramic radiography is commonly used for radiological examinations before third molar extraction but has limitations in assessing the true contact relationship between the third molars and the mandibular canal. Therefore, the true relationship between the mandibular canal and molars can be determined only through additional cone-beam computed tomography (CBCT) imaging. In this study, we aimed to develop an automatic diagnosis method based on a deep learning model that can determine the true proximity between the mandibular canal and third molars using only panoramic radiographs. A total of 901 third molars shown on panoramic radiographs were examined with CBCT imaging to ascertain whether true proximity existed between the mandibular canal and the third molar by two radiologists (450 molars: true contact, 451 molars: true non-contact). Three deep learning models (RetinaNet, YOLOv3, and EfficientDet) were developed, with performance metrics of accuracy, sensitivity, and specificity. EfficientDet showed the highest performance, with an accuracy of 78.65%, sensitivity of 82.02%, and specificity of 75.28%. The proposed deep learning method can be helpful when clinicians must evaluate the proximity of the mandibular canal and a third molar using only panoramic radiographs without CBCT.
Collapse
Affiliation(s)
- Kug Jin Jeon
- Department of Oral and Maxillofacial Radiology, Yonsei University College of Dentistry, 50-1 Yonsei-ro Seodaemun-gu, Seoul, 03722, Korea
| | - Hanseung Choi
- Department of Oral and Maxillofacial Radiology, Yonsei University College of Dentistry, 50-1 Yonsei-ro Seodaemun-gu, Seoul, 03722, Korea
| | - Chena Lee
- Department of Oral and Maxillofacial Radiology, Yonsei University College of Dentistry, 50-1 Yonsei-ro Seodaemun-gu, Seoul, 03722, Korea
| | - Sang-Sun Han
- Department of Oral and Maxillofacial Radiology, Yonsei University College of Dentistry, 50-1 Yonsei-ro Seodaemun-gu, Seoul, 03722, Korea.
| |
Collapse
|
20
|
Picoli FF, Fontenele RC, Van der Cruyssen F, Ahmadzai I, Trigeminal Nerve Injuries Research Group, Politis C, Silva MAG, Jacobs R. Risk assessment of inferior alveolar nerve injury after wisdom tooth removal using 3D AI-driven models: A within-patient study. J Dent 2023; 139:104765. [PMID: 38353315 DOI: 10.1016/j.jdent.2023.104765] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2023] [Revised: 10/10/2023] [Accepted: 10/26/2023] [Indexed: 02/16/2024] Open
Abstract
OBJECTIVE To compare a three-dimensional (3D) artificial intelligence (AI)- driven model with panoramic radiography (PANO) and cone-beam computed tomography (CBCT) in assessing the risk of inferior alveolar nerve (IAN) injury after mandibular wisdom tooth (M3M) removal through a within-patient controlled trial. METHODS From a database of 6,010 patients undergoing M3M surgery, 25 patients met the inclusion criteria of bilateral M3M removal with postoperative unilateral IAN injury. In this within-patient controlled trial, preoperative PANO and CBCT images were available, while 3D-AI models of the mandibular canal and teeth were generated from the CBCT images using the Virtual Patient Creator AI platform (Relu BV, Leuven, Belgium). Five examiners, who were blinded to surgical outcomes, assessed the imaging modalities and assigned scores indicating the risk level of IAN injury (high, medium, or low risk). Sensitivity, specificity, and area under receiver operating curve (AUC) for IAN risk assessment were calculated for each imaging modality. RESULTS For IAN injury risk assessment after M3M removal, sensitivity was 0.87 for 3D-AI, 0.89 for CBCT versus 0.73 for PANO. Furthermore, the AUC and specificity values were 0.63 and 0.39 for 3D-AI, 0.58 and 0.28 for CBCT, and 0.57 and 0.41 for PANO, respectively. There was no statistically significant difference (p>0.05) among the imaging modalities for any diagnostic parameters. CONCLUSION This within-patient controlled trial study revealed that risk assessment for IAN injury after M3M removal was rather similar for 3D-AI, PANO, and CBCT, with a sensitivity for injury prediction reaching up to 0.87 for 3D-AI and 0.89 for CBCT. CLINICAL SIGNIFICANCE This within-patient trial is pioneering in exploring the application of 3D AI-driven models for assessing IAN injury risk after M3M removal. The present results indicate that AI-powered 3D models based on CBCT might facilitate IAN risk assessment of M3M removal.
Collapse
Affiliation(s)
- Fernando Fortes Picoli
- OMFS IMPATH Research Group, Department of Imaging and Pathology, Faculty of Medicine, University of Leuven and Department of Oral & Maxillofacial Surgery, University Hospitals Leuven, KU Leuven, Kapucijnenvoer 7, 3000, Leuven, Belgium; School of Dentistry, Federal University of Goiás, Goiânia, GO, Brazil
| | - Rocharles Cavalcante Fontenele
- OMFS IMPATH Research Group, Department of Imaging and Pathology, Faculty of Medicine, University of Leuven and Department of Oral & Maxillofacial Surgery, University Hospitals Leuven, KU Leuven, Kapucijnenvoer 7, 3000, Leuven, Belgium; Department of Oral Diagnosis, Piracicaba Dental School, University of Campinas, Piracicaba, Sao Paulo, Brazil
| | - Frederic Van der Cruyssen
- OMFS IMPATH Research Group, Department of Imaging and Pathology, Faculty of Medicine, University of Leuven and Department of Oral & Maxillofacial Surgery, University Hospitals Leuven, KU Leuven, Kapucijnenvoer 7, 3000, Leuven, Belgium; Department of Oral and Maxillofacial Surgery, University Hospitals Leuven, Leuven, Belgium
| | - Iraj Ahmadzai
- OMFS IMPATH Research Group, Department of Imaging and Pathology, Faculty of Medicine, University of Leuven and Department of Oral & Maxillofacial Surgery, University Hospitals Leuven, KU Leuven, Kapucijnenvoer 7, 3000, Leuven, Belgium
| | | | - Constantinus Politis
- OMFS IMPATH Research Group, Department of Imaging and Pathology, Faculty of Medicine, University of Leuven and Department of Oral & Maxillofacial Surgery, University Hospitals Leuven, KU Leuven, Kapucijnenvoer 7, 3000, Leuven, Belgium; Department of Oral and Maxillofacial Surgery, University Hospitals Leuven, Leuven, Belgium
| | | | - Reinhilde Jacobs
- OMFS IMPATH Research Group, Department of Imaging and Pathology, Faculty of Medicine, University of Leuven and Department of Oral & Maxillofacial Surgery, University Hospitals Leuven, KU Leuven, Kapucijnenvoer 7, 3000, Leuven, Belgium; Department of Dental Medicine, Karolinska Institutet, Stockholm, Sweden.
| |
Collapse
|
21
|
Carvalho J, Lotz M, Rubi L, Unger S, Pfister T, Buhmann J, Stadlinger B. Preinterventional Third-Molar Assessment Using Robust Machine Learning. J Dent Res 2023; 102:1452-1459. [PMID: 37944556 PMCID: PMC10683342 DOI: 10.1177/00220345231200786] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2023] Open
Abstract
Machine learning (ML) models, especially deep neural networks, are increasingly being used for the analysis of medical images and as a supporting tool for clinical decision-making. In this study, we propose an artificial intelligence system to facilitate dental decision-making for the removal of mandibular third molars (M3M) based on 2-dimensional orthopantograms and the risk assessment of such a procedure. A total of 4,516 panoramic radiographic images collected at the Center of Dental Medicine at the University of Zurich, Switzerland, were used for training the ML model. After image preparation and preprocessing, a spatially dependent U-Net was employed to detect and retrieve the region of the M3M and inferior alveolar nerve (IAN). Image patches identified to contain a M3M were automatically processed by a deep neural network for the classification of M3M superimposition over the IAN (task 1) and M3M root development (task 2). A control evaluation set of 120 images, collected from a different data source than the training data and labeled by 5 dental practitioners, was leveraged to reliably evaluate model performance. By 10-fold cross-validation, we achieved accuracy values of 0.94 and 0.93 for the M3M-IAN superimposition task and the M3M root development task, respectively, and accuracies of 0.9 and 0.87 when evaluated on the control data set, using a ResNet-101 trained in a semisupervised fashion. Matthew's correlation coefficient values of 0.82 and 0.75 for task 1 and task 2, evaluated on the control data set, indicate robust generalization of our model. Depending on the different label combinations of task 1 and task 2, we propose a diagnostic table that suggests whether additional imaging via 3-dimensional cone beam tomography is advisable. Ultimately, computer-aided decision-making tools benefit clinical practice by enabling efficient and risk-reduced decision-making and by supporting less experienced practitioners before the surgical removal of the M3M.
Collapse
Affiliation(s)
- J.S. Carvalho
- ETH Zurich, Department of Computer Science, Zurich, Switzerland
- ETH AI Center, Zurich, Switzerland
| | - M. Lotz
- University of Zurich, Center for Dental Medicine, Zurich, Switzerland
| | - L. Rubi
- ETH Zurich, Department of Computer Science, Zurich, Switzerland
| | - S. Unger
- University of Zurich, Center for Dental Medicine, Zurich, Switzerland
| | - T. Pfister
- University of Zurich, Center for Dental Medicine, Zurich, Switzerland
| | - J.M. Buhmann
- ETH Zurich, Department of Computer Science, Zurich, Switzerland
- ETH AI Center, Zurich, Switzerland
| | - B. Stadlinger
- University of Zurich, Center for Dental Medicine, Zurich, Switzerland
- ETH AI Center, Zurich, Switzerland
| |
Collapse
|
22
|
Chun SY, Kang YH, Yang S, Kang SR, Lee SJ, Kim JM, Kim JE, Huh KH, Lee SS, Heo MS, Yi WJ. Automatic classification of 3D positional relationship between mandibular third molar and inferior alveolar canal using a distance-aware network. BMC Oral Health 2023; 23:794. [PMID: 37880603 PMCID: PMC10598947 DOI: 10.1186/s12903-023-03496-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2023] [Accepted: 10/05/2023] [Indexed: 10/27/2023] Open
Abstract
The purpose of this study was to automatically classify the three-dimensional (3D) positional relationship between an impacted mandibular third molar (M3) and the inferior alveolar canal (MC) using a distance-aware network in cone-beam CT (CBCT) images. We developed a network consisting of cascaded stages of segmentation and classification for the buccal-lingual relationship between the M3 and the MC. The M3 and the MC were simultaneously segmented using Dense121 U-Net in the segmentation stage, and their buccal-lingual relationship was automatically classified using a 3D distance-aware network with the multichannel inputs of the original CBCT image and the signed distance map (SDM) generated from the segmentation in the classification stage. The Dense121 U-Net achieved the highest average precision of 0.87, 0.96, and 0.94 in the segmentation of the M3, the MC, and both together, respectively. The 3D distance-aware classification network of the Dense121 U-Net with the input of both the CBCT image and the SDM showed the highest performance of accuracy, sensitivity, specificity, and area under the receiver operating characteristic curve, each of which had a value of 1.00. The SDM generated from the segmentation mask significantly contributed to increasing the accuracy of the classification network. The proposed distance-aware network demonstrated high accuracy in the automatic classification of the 3D positional relationship between the M3 and the MC by learning anatomical and geometrical information from the CBCT images.
Collapse
Affiliation(s)
- So-Young Chun
- Interdisciplinary Program in Bioengineering, Graduate School of Engineering, Seoul National University, Seoul, South Korea
| | - Yun-Hui Kang
- Department of Oral and Maxillofacial Radiology, Seoul National University Dental Hospital, Seoul, South Korea
| | - Su Yang
- Department of Applied Bioengineering, Graduate School of Convergence Science and Technology, Seoul National University, Seoul, South Korea
| | - Se-Ryong Kang
- Department of Applied Bioengineering, Graduate School of Convergence Science and Technology, Seoul National University, Seoul, South Korea
| | | | - Jun-Min Kim
- Department of Electronics and Information Engineering, Hansung University, Seoul, South Korea
| | - Jo-Eun Kim
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, Seoul, South Korea
| | - Kyung-Hoe Huh
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, Seoul, South Korea
| | - Sam-Sun Lee
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, Seoul, South Korea
| | - Min-Suk Heo
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, Seoul, South Korea
| | - Won-Jin Yi
- Interdisciplinary Program in Bioengineering, Graduate School of Engineering, Seoul National University, Seoul, South Korea.
- Department of Applied Bioengineering, Graduate School of Convergence Science and Technology, Seoul National University, Seoul, South Korea.
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, Seoul, South Korea.
| |
Collapse
|
23
|
Fan W, Zhang J, Wang N, Li J, Hu L. The Application of Deep Learning on CBCT in Dentistry. Diagnostics (Basel) 2023; 13:2056. [PMID: 37370951 DOI: 10.3390/diagnostics13122056] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2023] [Revised: 06/06/2023] [Accepted: 06/12/2023] [Indexed: 06/29/2023] Open
Abstract
Cone beam computed tomography (CBCT) has become an essential tool in modern dentistry, allowing dentists to analyze the relationship between teeth and the surrounding tissues. However, traditional manual analysis can be time-consuming and its accuracy depends on the user's proficiency. To address these limitations, deep learning (DL) systems have been integrated into CBCT analysis to improve accuracy and efficiency. Numerous DL models have been developed for tasks such as automatic diagnosis, segmentation, classification of teeth, inferior alveolar nerve, bone, airway, and preoperative planning. All research articles summarized were from Pubmed, IEEE, Google Scholar, and Web of Science up to December 2022. Many studies have demonstrated that the application of deep learning technology in CBCT examination in dentistry has achieved significant progress, and its accuracy in radiology image analysis has reached the level of clinicians. However, in some fields, its accuracy still needs to be improved. Furthermore, ethical issues and CBCT device differences may prohibit its extensive use. DL models have the potential to be used clinically as medical decision-making aids. The combination of DL and CBCT can highly reduce the workload of image reading. This review provides an up-to-date overview of the current applications of DL on CBCT images in dentistry, highlighting its potential and suggesting directions for future research.
Collapse
Affiliation(s)
- Wenjie Fan
- Department of Stomatology, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan 430022, China
- School of Stomatology, Tongji Medical College, Huazhong University of Science and Technology, Wuhan 430030, China
| | - Jiaqi Zhang
- Department of Stomatology, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan 430022, China
- School of Stomatology, Tongji Medical College, Huazhong University of Science and Technology, Wuhan 430030, China
| | - Nan Wang
- Department of Stomatology, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan 430022, China
- School of Stomatology, Tongji Medical College, Huazhong University of Science and Technology, Wuhan 430030, China
| | - Jia Li
- Department of Stomatology, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan 430022, China
- School of Stomatology, Tongji Medical College, Huazhong University of Science and Technology, Wuhan 430030, China
| | - Li Hu
- Department of Stomatology, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan 430022, China
- School of Stomatology, Tongji Medical College, Huazhong University of Science and Technology, Wuhan 430030, China
| |
Collapse
|
24
|
Yang P, Guo X, Mu C, Qi S, Li G. Detection of vertical root fractures by cone-beam computed tomography based on deep learning. Dentomaxillofac Radiol 2023; 52:20220345. [PMID: 36802858 PMCID: PMC9944014 DOI: 10.1259/dmfr.20220345] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2022] [Revised: 01/31/2023] [Accepted: 01/31/2023] [Indexed: 02/23/2023] Open
Abstract
OBJECTIVES This study aims to evaluate the performance of ResNet models in the detection of in vitro and in vivo vertical root fractures (VRF) in Cone-beam Computed Tomography (CBCT) images. METHODS A CBCT image dataset consisting of 28 teeth (14 intact and 14 teeth with VRF, 1641 slices) from 14 patients, and another dataset containing 60 teeth (30 intact and 30 teeth with VRF, 3665 slices) from an in vitro model were used for the establishment of VRFconvolutional neural network (CNN) models. The most popular CNN architecture ResNet with different layers was fine-tuned for the detection of VRF. Sensitivity, specificity, accuracy, PPV (positive predictive value), NPV (negative predictive value), and AUC (the area under the receiver operating characteristic curve) of the VRF slices classified by the CNN in the test set were compared. Two oral and maxillofacial radiologists independently reviewed all the CBCT images of the test set, and intraclass correlation coefficients (ICCs) were calculated to assess the interobserver agreement for the oral maxillofacial radiologists. RESULTS The AUC of the models on the patient data were: 0.827(ResNet-18), 0.929(ResNet-50), and 0.882(ResNet-101). The AUC of the models on the mixed data get improved as:0.927(ResNet-18), 0.936(ResNet-50), and 0.893(ResNet-101). The maximum AUC were: 0.929 (0.908-0.950, 95% CI) and 0.936 (0.924-0.948, 95% CI) for the patient data and mixed data from ResNet-50, which is comparable to the AUC (0.937 and 0.950) for patient data and (0.915 and 0.935) for the mixed data obtained from the two oral and maxillofacial radiologists, respectively. CONCLUSIONS Deep-learning models showed high accuracy in the detection of VRF using CBCT images. The data obtained from the in vitro VRF model increases the data scale, which is beneficial to the training of deep-learning models.
Collapse
Affiliation(s)
| | | | | | - Senrong Qi
- Department of Oral and Maxillofacial Radiology, Beijing Stomatology Hospital, School of Stomatology, Capital Medical University, Beijing, China
| | - Gang Li
- Department of Oral and Maxillofacial Radiology, Peking University School and Hospital of Stomatology & National Center of Stomatology & National Clinical Research Center for Oral Diseases & National Engineering Laboratory for Digital and Material Technology of Stomatology & Beijing Key Laboratory of Digital Stomatology & Research Center of Engineering and Technology for Computerized Dentistry Ministry of Health & NMPA Key Laboratory for Dental Materials, Beijing, China
| |
Collapse
|
25
|
Evaluation of the Diagnostic and Prognostic Accuracy of Artificial Intelligence in Endodontic Dentistry: A Comprehensive Review of Literature. COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE 2023; 2023:7049360. [PMID: 36761829 PMCID: PMC9904932 DOI: 10.1155/2023/7049360] [Citation(s) in RCA: 16] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/09/2021] [Revised: 10/23/2022] [Accepted: 11/26/2022] [Indexed: 02/01/2023]
Abstract
Aim This comprehensive review is aimed at evaluating the diagnostic and prognostic accuracy of artificial intelligence in endodontic dentistry. Introduction Artificial intelligence (AI) is a relatively new technology that has widespread use in dentistry. The AI technologies have primarily been used in dentistry to diagnose dental diseases, plan treatment, make clinical decisions, and predict the prognosis. AI models like convolutional neural networks (CNN) and artificial neural networks (ANN) have been used in endodontics to study root canal system anatomy, determine working length measurements, detect periapical lesions and root fractures, predict the success of retreatment procedures, and predict the viability of dental pulp stem cells. Methodology. The literature was searched in electronic databases such as Google Scholar, Medline, PubMed, Embase, Web of Science, and Scopus, published over the last four decades (January 1980 to September 15, 2021) by using keywords such as artificial intelligence, machine learning, deep learning, application, endodontics, and dentistry. Results The preliminary search yielded 2560 articles relevant enough to the paper's purpose. A total of 88 articles met the eligibility criteria. The majority of research on AI application in endodontics has concentrated on tracing apical foramen, verifying the working length, projection of periapical pathologies, root morphologies, and retreatment predictions and discovering the vertical root fractures. Conclusion In endodontics, AI displayed accuracy in terms of diagnostic and prognostic evaluations. The use of AI can help enhance the treatment plan, which in turn can lead to an increase in the success rate of endodontic treatment outcomes. The AI is used extensively in endodontics and could help in clinical applications, such as detecting root fractures, periapical pathologies, determining working length, tracing apical foramen, the morphology of root, and disease prediction.
Collapse
|
26
|
Hung KF, Yeung AWK, Bornstein MM, Schwendicke F. Personalized dental medicine, artificial intelligence, and their relevance for dentomaxillofacial imaging. Dentomaxillofac Radiol 2023; 52:20220335. [PMID: 36472627 PMCID: PMC9793453 DOI: 10.1259/dmfr.20220335] [Citation(s) in RCA: 27] [Impact Index Per Article: 13.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/15/2022] [Revised: 11/08/2022] [Accepted: 11/11/2022] [Indexed: 12/12/2022] Open
Abstract
Personalized medicine refers to the tailoring of diagnostics and therapeutics to individuals based on one's biological, social, and behavioral characteristics. While personalized dental medicine is still far from being a reality, advanced artificial intelligence (AI) technologies with improved data analytic approaches are expected to integrate diverse data from the individual, setting, and system levels, which may facilitate a deeper understanding of the interaction of these multilevel data and therefore bring us closer to more personalized, predictive, preventive, and participatory dentistry, also known as P4 dentistry. In the field of dentomaxillofacial imaging, a wide range of AI applications, including several commercially available software options, have been proposed to assist dentists in the diagnosis and treatment planning of various dentomaxillofacial diseases, with performance similar or even superior to that of specialists. Notably, the impact of these dental AI applications on treatment decision, clinical and patient-reported outcomes, and cost-effectiveness has so far been assessed sparsely. Such information should be further investigated in future studies to provide patients, providers, and healthcare organizers a clearer picture of the true usefulness of AI in daily dental practice.
Collapse
Affiliation(s)
- Kuo Feng Hung
- Division of Oral and Maxillofacial Surgery, Faculty of Dentistry, The University of Hong Kong, Hong Kong SAR, China
| | - Andy Wai Kan Yeung
- Division of Oral and Maxillofacial Radiology, Applied Oral Sciences and Community Dental Care, Faculty of Dentistry, The University of Hong Kong, Hong Kong SAR, China
| | - Michael M. Bornstein
- Department of Oral Health & Medicine, University Center for Dental Medicine Basel UZB, University of Basel, Basel, Switzerland
| | - Falk Schwendicke
- Department of Oral Diagnostics, Digital Health and Health Services Research, Charité–Universitätsmedizin Berlin, Berlin, Germany
| |
Collapse
|
27
|
Jiang Y, Shang F, Peng J, Liang J, Fan Y, Yang Z, Qi Y, Yang Y, Xu T, Jiang R. Automatic Masseter Muscle Accurate Segmentation from CBCT Using Deep Learning-Based Model. J Clin Med 2022; 12:jcm12010055. [PMID: 36614860 PMCID: PMC9820952 DOI: 10.3390/jcm12010055] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2022] [Revised: 12/17/2022] [Accepted: 12/18/2022] [Indexed: 12/24/2022] Open
Abstract
Segmentation of the masseter muscle (MM) on cone-beam computed tomography (CBCT) is challenging due to the lack of sufficient soft-tissue contrast. Moreover, manual segmentation is laborious and time-consuming. The purpose of this study was to propose a deep learning-based automatic approach to accurately segment the MM from CBCT under the refinement of high-quality paired computed tomography (CT). Fifty independent CBCT and 42 clinically hard-to-obtain paired CBCT and CT were manually annotated by two observers. A 3D U-shape network was carefully designed to segment the MM effectively. Manual annotations on CT were set as the ground truth. Additionally, an extra five CT and five CBCT auto-segmentation results were revised by one oral and maxillofacial anatomy expert to evaluate their clinical suitability. CBCT auto-segmentation results were comparable to the CT counterparts and significantly improved the similarity with the ground truth compared with manual annotations on CBCT. The automatic approach was more than 332 times shorter than that of a human operation. Only 0.52% of the manual revision fraction was required. This automatic model could simultaneously and accurately segment the MM structures on CBCT and CT, which can improve clinical efficiency and efficacy, and provide critical information for personalized treatment and long-term follow-up.
Collapse
Affiliation(s)
- Yiran Jiang
- Department of Orthodontics, Peking University School and Hospital of Stomatology, Beijing 100081, China
- National Clinical Research Center for Oral Diseases, Beijing 100081, China
- National Engineering Laboratory for Digital and Material Technology of Stomatology, Beijing Key Laboratory of Digital Stomatology, Peking University School and Hospital of Stomatology, Beijing 100081, China
- NHC Research Center of Engineering and Technology for Computerized Dentistry, Beijing 100081, China
| | - Fangxin Shang
- Intelligent Healthcare Unit, Baidu, Beijing 100081, China
| | - Jiale Peng
- Department of Orthodontics, Peking University School and Hospital of Stomatology, Beijing 100081, China
- National Clinical Research Center for Oral Diseases, Beijing 100081, China
- National Engineering Laboratory for Digital and Material Technology of Stomatology, Beijing Key Laboratory of Digital Stomatology, Peking University School and Hospital of Stomatology, Beijing 100081, China
- NHC Research Center of Engineering and Technology for Computerized Dentistry, Beijing 100081, China
| | - Jie Liang
- National Clinical Research Center for Oral Diseases, Beijing 100081, China
- National Engineering Laboratory for Digital and Material Technology of Stomatology, Beijing Key Laboratory of Digital Stomatology, Peking University School and Hospital of Stomatology, Beijing 100081, China
- NHC Research Center of Engineering and Technology for Computerized Dentistry, Beijing 100081, China
- Department of Oral and Maxillofacial Surgery, Peking University School and Hospital of Stomatology, Beijing 100081, China
| | - Yi Fan
- Department of Orthodontics, Peking University School and Hospital of Stomatology, Beijing 100081, China
- National Clinical Research Center for Oral Diseases, Beijing 100081, China
- National Engineering Laboratory for Digital and Material Technology of Stomatology, Beijing Key Laboratory of Digital Stomatology, Peking University School and Hospital of Stomatology, Beijing 100081, China
- NHC Research Center of Engineering and Technology for Computerized Dentistry, Beijing 100081, China
| | - Zhongpeng Yang
- Department of Orthodontics, Peking University School and Hospital of Stomatology, Beijing 100081, China
- National Clinical Research Center for Oral Diseases, Beijing 100081, China
- National Engineering Laboratory for Digital and Material Technology of Stomatology, Beijing Key Laboratory of Digital Stomatology, Peking University School and Hospital of Stomatology, Beijing 100081, China
- NHC Research Center of Engineering and Technology for Computerized Dentistry, Beijing 100081, China
| | - Yuhan Qi
- Department of Orthodontics, Peking University School and Hospital of Stomatology, Beijing 100081, China
- National Clinical Research Center for Oral Diseases, Beijing 100081, China
- National Engineering Laboratory for Digital and Material Technology of Stomatology, Beijing Key Laboratory of Digital Stomatology, Peking University School and Hospital of Stomatology, Beijing 100081, China
- NHC Research Center of Engineering and Technology for Computerized Dentistry, Beijing 100081, China
| | - Yehui Yang
- Intelligent Healthcare Unit, Baidu, Beijing 100081, China
| | - Tianmin Xu
- Department of Orthodontics, Peking University School and Hospital of Stomatology, Beijing 100081, China
- National Clinical Research Center for Oral Diseases, Beijing 100081, China
- National Engineering Laboratory for Digital and Material Technology of Stomatology, Beijing Key Laboratory of Digital Stomatology, Peking University School and Hospital of Stomatology, Beijing 100081, China
- NHC Research Center of Engineering and Technology for Computerized Dentistry, Beijing 100081, China
- Correspondence: (T.X.); (R.J.); Tel.: +86-10-8219-5330 (T.X.); +86-10-8129-5737 (R.J.)
| | - Ruoping Jiang
- Department of Orthodontics, Peking University School and Hospital of Stomatology, Beijing 100081, China
- National Clinical Research Center for Oral Diseases, Beijing 100081, China
- National Engineering Laboratory for Digital and Material Technology of Stomatology, Beijing Key Laboratory of Digital Stomatology, Peking University School and Hospital of Stomatology, Beijing 100081, China
- NHC Research Center of Engineering and Technology for Computerized Dentistry, Beijing 100081, China
- Correspondence: (T.X.); (R.J.); Tel.: +86-10-8219-5330 (T.X.); +86-10-8129-5737 (R.J.)
| |
Collapse
|
28
|
Huang Z, Zheng H, Huang J, Yang Y, Wu Y, Ge L, Wang L. The Construction and Evaluation of a Multi-Task Convolutional Neural Network for a Cone-Beam Computed-Tomography-Based Assessment of Implant Stability. Diagnostics (Basel) 2022; 12:2673. [PMID: 36359516 PMCID: PMC9689694 DOI: 10.3390/diagnostics12112673] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/08/2022] [Revised: 10/21/2022] [Accepted: 10/29/2022] [Indexed: 07/21/2023] Open
Abstract
Objectives: Assessing implant stability is integral to dental implant therapy. This study aimed to construct a multi-task cascade convolution neural network to evaluate implant stability using cone-beam computed tomography (CBCT). Methods: A dataset of 779 implant coronal section images was obtained from CBCT scans, and matching clinical information was used for the training and test datasets. We developed a multi-task cascade network based on CBCT to assess implant stability. We used the MobilenetV2-DeeplabV3+ semantic segmentation network, combined with an image processing algorithm in conjunction with prior knowledge, to generate the volume of interest (VOI) that was eventually used for the ResNet-50 classification of implant stability. The performance of the multitask cascade network was evaluated in a test set by comparing the implant stability quotient (ISQ), measured using an Osstell device. Results: The cascade network established in this study showed good prediction performance for implant stability classification. The binary, ternary, and quaternary ISQ classification test set accuracies were 96.13%, 95.33%, and 92.90%, with mean precisions of 96.20%, 95.33%, and 93.71%, respectively. In addition, this cascade network evaluated each implant's stability in only 3.76 s, indicating high efficiency. Conclusions: To our knowledge, this is the first study to present a CBCT-based deep learning approach CBCT to assess implant stability. The multi-task cascade network accomplishes a series of tasks related to implant denture segmentation, VOI extraction, and implant stability classification, and has good concordance with the ISQ.
Collapse
Affiliation(s)
- Zelun Huang
- Guangzhou Key Laboratory of Basic and Applied Research of Oral Regenerative Medicine, Guangdong Engineering Research Center of Oral Restoration and Reconstruction, Affiliated Stomatology Hospital of Guangzhou Medical University, Guangzhou 510182, China
| | - Haoran Zheng
- Department of Chemical & Materials Engineering, University of Auckland, Auckland 1010, New Zealand
| | - Junqiang Huang
- Department of Stomatology, The Third Affiliated Hospital of Guangzhou Medical University, Guangzhou 510150, China
| | - Yang Yang
- Guangzhou Key Laboratory of Basic and Applied Research of Oral Regenerative Medicine, Guangdong Engineering Research Center of Oral Restoration and Reconstruction, Affiliated Stomatology Hospital of Guangzhou Medical University, Guangzhou 510182, China
| | - Yupeng Wu
- Guangzhou Key Laboratory of Basic and Applied Research of Oral Regenerative Medicine, Guangdong Engineering Research Center of Oral Restoration and Reconstruction, Affiliated Stomatology Hospital of Guangzhou Medical University, Guangzhou 510182, China
| | - Linhu Ge
- Guangzhou Key Laboratory of Basic and Applied Research of Oral Regenerative Medicine, Guangdong Engineering Research Center of Oral Restoration and Reconstruction, Affiliated Stomatology Hospital of Guangzhou Medical University, Guangzhou 510182, China
| | - Liping Wang
- Guangzhou Key Laboratory of Basic and Applied Research of Oral Regenerative Medicine, Guangdong Engineering Research Center of Oral Restoration and Reconstruction, Affiliated Stomatology Hospital of Guangzhou Medical University, Guangzhou 510182, China
| |
Collapse
|
29
|
Analysis of Deep Learning Techniques for Dental Informatics: A Systematic Literature Review. Healthcare (Basel) 2022; 10:healthcare10101892. [PMID: 36292339 PMCID: PMC9602147 DOI: 10.3390/healthcare10101892] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/31/2022] [Revised: 08/30/2022] [Accepted: 08/31/2022] [Indexed: 12/04/2022] Open
Abstract
Within the ever-growing healthcare industry, dental informatics is a burgeoning field of study. One of the major obstacles to the health care system’s transformation is obtaining knowledge and insightful data from complex, high-dimensional, and diverse sources. Modern biomedical research, for instance, has seen an increase in the use of complex, heterogeneous, poorly documented, and generally unstructured electronic health records, imaging, sensor data, and text. There were still certain restrictions even after many current techniques were used to extract more robust and useful elements from the data for analysis. New effective paradigms for building end-to-end learning models from complex data are provided by the most recent deep learning technology breakthroughs. Therefore, the current study aims to examine the most recent research on the use of deep learning techniques for dental informatics problems and recommend creating comprehensive and meaningful interpretable structures that might benefit the healthcare industry. We also draw attention to some drawbacks and the need for better technique development and provide new perspectives about this exciting new development in the field.
Collapse
|
30
|
A Fused Deep Learning Architecture for the Detection of the Relationship between the Mandibular Third Molar and the Mandibular Canal. Diagnostics (Basel) 2022; 12:diagnostics12082018. [PMID: 36010368 PMCID: PMC9407570 DOI: 10.3390/diagnostics12082018] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2022] [Revised: 08/17/2022] [Accepted: 08/18/2022] [Indexed: 12/01/2022] Open
Abstract
The study aimed to generate a fused deep learning algorithm that detects and classifies the relationship between the mandibular third molar and mandibular canal on orthopantomographs. Radiographs (n = 1880) were randomly selected from the hospital archive. Two dentomaxillofacial radiologists annotated the data via MATLAB and classified them into four groups according to the overlap of the root of the mandibular third molar and mandibular canal. Each radiograph was segmented using a U-Net-like architecture. The segmented images were classified by AlexNet. Accuracy, the weighted intersection over union score, the dice coefficient, specificity, sensitivity, and area under curve metrics were used to quantify the performance of the models. Also, three dental practitioners were asked to classify the same test data, their success rate was assessed using the Intraclass Correlation Coefficient. The segmentation network achieved a global accuracy of 0.99 and a weighted intersection over union score of 0.98, average dice score overall images was 0.91. The classification network achieved an accuracy of 0.80, per class sensitivity of 0.74, 0.83, 0.86, 0.67, per class specificity of 0.92, 0.95, 0.88, 0.96 and AUC score of 0.85. The most successful dental practitioner achieved a success rate of 0.79. The fused segmentation and classification networks produced encouraging results. The final model achieved almost the same classification performance as dental practitioners. Better diagnostic accuracy of the combined artificial intelligence tools may help to improve the prediction of the risk factors, especially for recognizing such anatomical variations.
Collapse
|
31
|
Potential and impact of artificial intelligence algorithms in dento-maxillofacial radiology. Clin Oral Investig 2022; 26:5535-5555. [PMID: 35438326 DOI: 10.1007/s00784-022-04477-y] [Citation(s) in RCA: 20] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2021] [Accepted: 03/25/2022] [Indexed: 12/20/2022]
Abstract
OBJECTIVES Novel artificial intelligence (AI) learning algorithms in dento-maxillofacial radiology (DMFR) are continuously being developed and improved using advanced convolutional neural networks. This review provides an overview of the potential and impact of AI algorithms in DMFR. MATERIALS AND METHODS A narrative review was conducted on the literature on AI algorithms in DMFR. RESULTS In the field of DMFR, AI algorithms were mainly proposed for (1) automated detection of dental caries, periapical pathologies, root fracture, periodontal/peri-implant bone loss, and maxillofacial cysts/tumors; (2) classification of mandibular third molars, skeletal malocclusion, and dental implant systems; (3) localization of cephalometric landmarks; and (4) improvement of image quality. Data insufficiency, overfitting, and the lack of interpretability are the main issues in the development and use of image-based AI algorithms. Several strategies have been suggested to address these issues, such as data augmentation, transfer learning, semi-supervised training, few-shot learning, and gradient-weighted class activation mapping. CONCLUSIONS Further integration of relevant AI algorithms into one fully automatic end-to-end intelligent system for possible multi-disciplinary applications is very likely to be a field of increased interest in the future. CLINICAL RELEVANCE This review provides dental practitioners and researchers with a comprehensive understanding of the current development, performance, issues, and prospects of image-based AI algorithms in DMFR.
Collapse
|
32
|
Issa J, Olszewski R, Dyszkiewicz-Konwińska M. The Effectiveness of Semi-Automated and Fully Automatic Segmentation for Inferior Alveolar Canal Localization on CBCT Scans: A Systematic Review. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2022; 19:560. [PMID: 35010820 PMCID: PMC8744855 DOI: 10.3390/ijerph19010560] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/21/2021] [Revised: 12/28/2021] [Accepted: 01/03/2022] [Indexed: 11/17/2022]
Abstract
This systematic review aims to identify the available semi-automatic and fully automatic algorithms for inferior alveolar canal localization as well as to present their diagnostic accuracy. Articles related to inferior alveolar nerve/canal localization using methods based on artificial intelligence (semi-automated and fully automated) were collected electronically from five different databases (PubMed, Medline, Web of Science, Cochrane, and Scopus). Two independent reviewers screened the titles and abstracts of the collected data, stored in EndnoteX7, against the inclusion criteria. Afterward, the included articles have been critically appraised to assess the quality of the studies using the Quality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2) tool. Seven studies were included following the deduplication and screening against exclusion criteria of the 990 initially collected articles. In total, 1288 human cone-beam computed tomography (CBCT) scans were investigated for inferior alveolar canal localization using different algorithms and compared to the results obtained from manual tracing executed by experts in the field. The reported values for diagnostic accuracy of the used algorithms were extracted. A wide range of testing measures was implemented in the analyzed studies, while some of the expected indexes were still missing in the results. Future studies should consider the new artificial intelligence guidelines to ensure proper methodology, reporting, results, and validation.
Collapse
Affiliation(s)
- Julien Issa
- Department of Biomaterials and Experimental Dentistry, Poznań University of Medical Sciences, Bukowska 70, 60-812 Poznań, Poland;
| | - Raphael Olszewski
- Department of Oral and Maxilofacial Surgery, Cliniques Universitaires Saint Luc, UCLouvain, Av. Hippocrate 10, 1200 Brussels, Belgium;
- Oral and Maxillofacial Surgery Research Lab (OMFS Lab), NMSK, Institut de Recherche Experimentale et Clinique, UCLouvain, Louvain-la-Neuve, 1348 Brussels, Belgium
| | - Marta Dyszkiewicz-Konwińska
- Department of Biomaterials and Experimental Dentistry, Poznań University of Medical Sciences, Bukowska 70, 60-812 Poznań, Poland;
| |
Collapse
|