Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 40
Filtrar
1.
J Oral Rehabil ; 2024 May 17.
Artigo em Inglês | MEDLINE | ID: mdl-38757865

RESUMO

BACKGROUND AND OBJECTIVE: The accurate diagnosis of temporomandibular disorders continues to be a challenge, despite the existence of internationally agreed-upon diagnostic criteria. The purpose of this study is to review applications of deep learning models in the diagnosis of temporomandibular joint arthropathies. MATERIALS AND METHODS: An electronic search was conducted on PubMed, Scopus, Embase, Google Scholar, IEEE, arXiv, and medRxiv up to June 2023. Studies that reported the efficacy (outcome) of prediction, object detection or classification of TMJ arthropathies by deep learning models (intervention) of human joint-based or arthrogenous TMDs (population) in comparison to reference standard (comparison) were included. To evaluate the risk of bias, included studies were critically analysed using the quality assessment of diagnostic accuracy studies (QUADAS-2). Diagnostic odds ratios (DOR) were calculated. Forrest plot and funnel plot were created using STATA 17 and MetaDiSc. RESULTS: Full text review was performed on 46 out of the 1056 identified studies and 21 studies met the eligibility criteria and were included in the systematic review. Four studies were graded as having a low risk of bias for all domains of QUADAS-2. The accuracy of all included studies ranged from 74% to 100%. Sensitivity ranged from 54% to 100%, specificity: 85%-100%, Dice coefficient: 85%-98%, and AUC: 77%-99%. The datasets were then pooled based on the sensitivity, specificity, and dataset size of seven studies that qualified for meta-analysis. The pooled sensitivity was 95% (85%-99%), specificity: 92% (86%-96%), and AUC: 97% (96%-98%). DORs were 232 (74-729). According to Deek's funnel plot and statistical evaluation (p =.49), publication bias was not present. CONCLUSION: Deep learning models can detect TMJ arthropathies high sensitivity and specificity. Clinicians, and especially those not specialized in orofacial pain, may benefit from this methodology for assessing TMD as it facilitates a rigorous and evidence-based framework, objective measurements, and advanced analysis techniques, ultimately enhancing diagnostic accuracy.

2.
BMC Oral Health ; 24(1): 574, 2024 May 17.
Artigo em Inglês | MEDLINE | ID: mdl-38760686

RESUMO

BACKGROUND: To develop and validate a deep learning model for automated assessment of endodontic case difficulty from periapical radiographs. METHODS: A dataset of 1,386 periapical radiographs was compiled from two clinical sites. Two dentists and two endodontists annotated the radiographs for difficulty using the "simple assessment" criteria from the American Association of Endodontists' case difficulty assessment form in the Endocase application. A classification task labeled cases as "easy" or "hard", while regression predicted overall difficulty scores. Convolutional neural networks (i.e. VGG16, ResNet18, ResNet50, ResNext50, and Inception v2) were used, with a baseline model trained via transfer learning from ImageNet weights. Other models was pre-trained using self-supervised contrastive learning (i.e. BYOL, SimCLR, MoCo, and DINO) on 20,295 unlabeled dental radiographs to learn representation without manual labels. Both models were evaluated using 10-fold cross-validation, with performance compared to seven human examiners (three general dentists and four endodontists) on a hold-out test set. RESULTS: The baseline VGG16 model attained 87.62% accuracy in classifying difficulty. Self-supervised pretraining did not improve performance. Regression predicted scores with ± 3.21 score error. All models outperformed human raters, with poor inter-examiner reliability. CONCLUSION: This pilot study demonstrated the feasibility of automated endodontic difficulty assessment via deep learning models.


Assuntos
Aprendizado Profundo , Humanos , Projetos Piloto , Radiografia Dentária , Redes Neurais de Computação
3.
Artigo em Inglês | MEDLINE | ID: mdl-38570273

RESUMO

OBJECTIVES: This study aims to evaluate the correctness of the generated answers by Google Bard, GPT-3.5, GPT-4, Claude-Instant, and Bing chatbots to decision-making clinical questions in the oral and maxillofacial surgery (OMFS) area. STUDY DESIGN: A group of 3 board-certified oral and maxillofacial surgeons designed a questionnaire with 50 case-based questions in multiple-choice and open-ended formats. Answers of chatbots to multiple-choice questions were examined against the chosen option by 3 referees. The chatbots' answers to the open-ended questions were evaluated based on the modified global quality scale. A P-value under .05 was considered significant. RESULTS: Bard, GPT-3.5, GPT-4, Claude-Instant, and Bing answered 34%, 36%, 38%, 38%, and 26% of the questions correctly, respectively. In open-ended questions, GPT-4 scored the most answers evaluated as grades "4" or "5," and Bing scored the most answers evaluated as grades "1" or "2." There were no statistically significant differences between the 5 chatbots in responding to the open-ended (P = .275) and multiple-choice (P = .699) questions. CONCLUSION: Considering the major inaccuracies in the responses of chatbots, despite their relatively good performance in answering open-ended questions, this technology yet cannot be trusted as a consultant for clinicians in decision-making situations.

4.
J Dent ; 144: 104938, 2024 May.
Artigo em Inglês | MEDLINE | ID: mdl-38499280

RESUMO

OBJECTIVES: Artificial Intelligence has applications such as Large Language Models (LLMs), which simulate human-like conversations. The potential of LLMs in healthcare is not fully evaluated. This pilot study assessed the accuracy and consistency of chatbots and clinicians in answering common questions in pediatric dentistry. METHODS: Two expert pediatric dentists developed thirty true or false questions involving different aspects of pediatric dentistry. Publicly accessible chatbots (Google Bard, ChatGPT4, ChatGPT 3.5, Llama, Sage, Claude 2 100k, Claude-instant, Claude-instant-100k, and Google Palm) were employed to answer the questions (3 independent new conversations). Three groups of clinicians (general dentists, pediatric specialists, and students; n = 20/group) also answered. Responses were graded by two pediatric dentistry faculty members, along with a third independent pediatric dentist. Resulting accuracies (percentage of correct responses) were compared using analysis of variance (ANOVA), and post-hoc pairwise group comparisons were corrected using Tukey's HSD method. ACronbach's alpha was calculated to determine consistency. RESULTS: Pediatric dentists were significantly more accurate (mean±SD 96.67 %± 4.3 %) than other clinicians and chatbots (p < 0.001). General dentists (88.0 % ± 6.1 %) also demonstrated significantly higher accuracy than chatbots (p < 0.001), followed by students (80.8 %±6.9 %). ChatGPT showed the highest accuracy (78 %±3 %) among chatbots. All chatbots except ChatGPT3.5 showed acceptable consistency (Cronbach alpha>0.7). CLINICAL SIGNIFICANCE: Based on this pilot study, chatbots may be valuable adjuncts for educational purposes and for distributing information to patients. However, they are not yet ready to serve as substitutes for human clinicians in diagnostic decision-making. CONCLUSION: In this pilot study, chatbots showed lower accuracy than dentists. Chatbots may not yet be recommended for clinical pediatric dentistry.


Assuntos
Odontólogos , Odontopediatria , Humanos , Projetos Piloto , Odontólogos/psicologia , Inteligência Artificial , Comunicação , Inquéritos e Questionários , Criança
5.
Artigo em Inglês | MEDLINE | ID: mdl-38553304

RESUMO

OBJECTIVES: In this study, we assessed 6 different artificial intelligence (AI) chatbots (Bing, GPT-3.5, GPT-4, Google Bard, Claude, Sage) responses to controversial and difficult questions in oral pathology, oral medicine, and oral radiology. STUDY DESIGN: The chatbots' answers were evaluated by board-certified specialists using a modified version of the global quality score on a 5-point Likert scale. The quality and validity of chatbot citations were evaluated. RESULTS: Claude had the highest mean score of 4.341 ± 0.582 for oral pathology and medicine. Bing had the lowest scores of 3.447 ± 0.566. In oral radiology, GPT-4 had the highest mean score of 3.621 ± 1.009 and Bing the lowest score of 2.379 ± 0.978. GPT-4 achieved the highest mean score of 4.066 ± 0.825 for performance across all disciplines. 82 out of 349 (23.50%) of generated citations from chatbots were fake. CONCLUSIONS: The most superior chatbot in providing high-quality information for controversial topics in various dental disciplines was GPT-4. Although the majority of chatbots performed well, it is suggested that developers of AI medical chatbots incorporate scientific citation authenticators to validate the outputted citations given the relatively high number of fabricated citations.


Assuntos
Inteligência Artificial , Medicina Bucal , Humanos , Radiologia , Patologia Bucal
6.
Pediatr Dent ; 46(1): 27-35, 2024 Jan 15.
Artigo em Inglês | MEDLINE | ID: mdl-38449036

RESUMO

Purpose: To systematically evaluate artificial intelligence applications for diagnostic and treatment planning possibilities in pediatric dentistry. Methods: PubMed®, EMBASE®, Scopus, Web of Science™, IEEE, medRxiv, arXiv, and Google Scholar were searched using specific search queries. The Quality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2) checklist was used to assess the risk of bias assessment of the included studies. Results: Based on the initial screening, 33 eligible studies were included (among 3,542). Eleven studies appeared to have low bias risk across all QUADAS-2 domains. Most applications focused on early childhood caries diagnosis and prediction, tooth identification, oral health evaluation, and supernumerary tooth identification. Six studies evaluated AI tools for mesiodens or supernumerary tooth identification on radigraphs, four for primary tooth identification and/or numbering, seven studies to detect caries on radiographs, and 12 to predict early childhood caries. For these four tasks, the reported accuracy of AI varied from 60 percent to 99 percent, sensitivity was from 20 percent to 100 percent, specificity was from 49 percent to 100 percent, F1-score was from 60 percent to 97 percent, and the area-under-the-curve varied from 87 percent to 100 percent. Conclusions: The overall body of evidence regarding artificial intelligence applications in pediatric dentistry does not allow for firm conclusions. For a wide range of applications, AI shows promising accuracy. Future studies should focus on a comparison of AI against the standard of care and employ a set of standardized outcomes and metrics to allow comparison across studies.


Assuntos
Inteligência Artificial , Odontopediatria , Criança , Pré-Escolar , Humanos , Cárie Dentária/diagnóstico por imagem , Cárie Dentária/terapia , Saúde Bucal , Dente Supranumerário
7.
J Endod ; 50(5): 562-578, 2024 May.
Artigo em Inglês | MEDLINE | ID: mdl-38387793

RESUMO

AIMS: The future dental and endodontic education must adapt to the current digitalized healthcare system in a hyper-connected world. The purpose of this scoping review was to investigate the ways an endodontic education curriculum could benefit from the implementation of artificial intelligence (AI) and overcome the limitations of this technology in the delivery of healthcare to patients. METHODS: An electronic search was carried out up to December 2023 using MEDLINE, Web of Science, Cochrane Library, and a manual search of reference literature. Grey literature, ongoing clinical trials were also searched using ClinicalTrials.gov. RESULTS: The search identified 251 records, of which 35 were deemed relevant to artificial intelligence (AI) and Endodontic education. Areas in which AI might aid students with their didactic and clinical endodontic education were identified as follows: 1) radiographic interpretation; 2) differential diagnosis; 3) treatment planning and decision-making; 4) case difficulty assessment; 5) preclinical training; 6) advanced clinical simulation and case-based training, 7) real-time clinical guidance; 8) autonomous systems and robotics; 9) progress evaluation and personalized education; 10) calibration and standardization. CONCLUSIONS: AI in endodontic education will support clinical and didactic teaching through individualized feedback; enhanced, augmented, and virtually generated training aids; automated detection and diagnosis; treatment planning and decision support; and AI-based student progress evaluation, and personalized education. Its implementation will inarguably change the current concept of teaching Endodontics. Dental educators would benefit from introducing AI in clinical and didactic pedagogy; however, they must be aware of AI's limitations and challenges to overcome.


Assuntos
Inteligência Artificial , Currículo , Educação em Odontologia , Endodontia , Endodontia/educação , Humanos , Educação em Odontologia/métodos , Competência Clínica
8.
Dentomaxillofac Radiol ; 53(1): 5-21, 2024 Jan 11.
Artigo em Inglês | MEDLINE | ID: mdl-38183164

RESUMO

OBJECTIVES: Improved tools based on deep learning can be used to accurately number and identify teeth. This study aims to review the use of deep learning in tooth numbering and identification. METHODS: An electronic search was performed through October 2023 on PubMed, Scopus, Cochrane, Google Scholar, IEEE, arXiv, and medRxiv. Studies that used deep learning models with segmentation, object detection, or classification tasks for teeth identification and numbering of human dental radiographs were included. For risk of bias assessment, included studies were critically analysed using quality assessment of diagnostic accuracy studies (QUADAS-2). To generate plots for meta-analysis, MetaDiSc and STATA 17 (StataCorp LP, College Station, TX, USA) were used. Pooled outcome diagnostic odds ratios (DORs) were determined through calculation. RESULTS: The initial search yielded 1618 studies, of which 29 were eligible based on the inclusion criteria. Five studies were found to have low bias across all domains of the QUADAS-2 tool. Deep learning has been reported to have an accuracy range of 81.8%-99% in tooth identification and numbering and a precision range of 84.5%-99.94%. Furthermore, sensitivity was reported as 82.7%-98% and F1-scores ranged from 87% to 98%. Sensitivity was 75.5%-98% and specificity was 79.9%-99%. Only 6 studies found the deep learning model to be less than 90% accurate. The average DOR of the pooled data set was 1612, the sensitivity was 89%, the specificity was 99%, and the area under the curve was 96%. CONCLUSION: Deep learning models successfully can detect, identify, and number teeth on dental radiographs. Deep learning-powered tooth numbering systems can enhance complex automated processes, such as accurately reporting which teeth have caries, thus aiding clinicians in making informed decisions during clinical practice.


Assuntos
Aprendizado Profundo , Cárie Dentária , Dente , Humanos , Radiografia Dentária , Dente/diagnóstico por imagem
9.
Clin Oral Investig ; 28(1): 88, 2024 Jan 13.
Artigo em Inglês | MEDLINE | ID: mdl-38217733

RESUMO

OBJECTIVE: This study aimed to review and synthesize studies using artificial intelligence (AI) for classifying, detecting, or segmenting oral mucosal lesions on photographs. MATERIALS AND METHOD: Inclusion criteria were (1) studies employing AI to (2) classify, detect, or segment oral mucosa lesions, (3) on oral photographs of human subjects. Included studies were assessed for risk of bias using Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2). A PubMed, Scopus, Embase, Web of Science, IEEE, arXiv, medRxiv, and grey literature (Google Scholar) search was conducted until June 2023, without language limitation. RESULTS: After initial searching, 36 eligible studies (from 8734 identified records) were included. Based on QUADAS-2, only 7% of studies were at low risk of bias for all domains. Studies employed different AI models and reported a wide range of outcomes and metrics. The accuracy of AI for detecting oral mucosal lesions ranged from 74 to 100%, while that for clinicians un-aided by AI ranged from 61 to 98%. Pooled diagnostic odds ratio for studies which evaluated AI for diagnosing or discriminating potentially malignant lesions was 155 (95% confidence interval 23-1019), while that for cancerous lesions was 114 (59-221). CONCLUSIONS: AI may assist in oral mucosa lesion screening while the expected accuracy gains or further health benefits remain unclear so far. CLINICAL RELEVANCE: Artificial intelligence assists oral mucosa lesion screening and may foster more targeted testing and referral in the hands of non-specialist providers, for example. So far, it remains unclear if accuracy gains compared with specialized can be realized.


Assuntos
Inteligência Artificial , Mucosa Bucal , Humanos , Encaminhamento e Consulta
10.
J Endod ; 50(2): 144-153.e2, 2024 Feb.
Artigo em Inglês | MEDLINE | ID: mdl-37977219

RESUMO

INTRODUCTION: The aim of this study was to leverage label-efficient self-supervised learning (SSL) to train a model that can detect ECR and differentiate it from caries. METHODS: Periapical (PA) radiographs of teeth with ECR defects were collected. Two board-certified endodontists reviewed PA radiographs and cone beam computed tomographic (CBCT) images independently to determine presence of ECR (ground truth). Radiographic data were divided into 3 regions of interest (ROIs): healthy teeth, teeth with ECR, and teeth with caries. Nine contrastive SSL models (SimCLR v2, MoCo v2, BYOL, DINO, NNCLR, SwAV, MSN, Barlow Twins, and SimSiam) were implemented in the assessment alongside 7 baseline deep learning models (ResNet-18, ResNet-50, VGG16, DenseNet, MobileNetV2, ResNeXt-50, and InceptionV3). A 10-fold cross-validation strategy and a hold-out test set were employed for model evaluation. Model performance was assessed via various metrics including classification accuracy, precision, recall, and F1-score. RESULTS: Included were 190 PA radiographs, composed of 470 ROIs. Results from 10-fold cross-validation demonstrated that most SSL models outperformed the transfer learning baseline models, with DINO achieving the highest mean accuracy (85.64 ± 4.56), significantly outperforming 13 other models (P < .05). DINO reached the highest test set (ie, 3 ROIs) accuracy (84.09%) while MoCo v2 exhibited the highest recall and F1-score (77.37% and 82.93%, respectively). CONCLUSIONS: This study showed that AI can assist clinicians in detecting ECR and differentiating it from caries. Additionally, it introduced the application of SSL in detecting ECR, emphasizing that SSL-based models can outperform transfer learning baselines and reduce reliance on large, labeled datasets.


Assuntos
Cárie Dentária , Dente , Humanos , Tomografia Computadorizada de Feixe Cônico/métodos , Inteligência Artificial , Tomografia Computadorizada por Raios X/métodos , Aprendizado de Máquina Supervisionado
11.
Oral Radiol ; 40(1): 1-20, 2024 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-37855976

RESUMO

PURPOSE: This study aims to review deep learning applications for detecting head and neck cancer (HNC) using magnetic resonance imaging (MRI) and radiographic data. METHODS: Through January 2023, a PubMed, Scopus, Embase, Google Scholar, IEEE, and arXiv search were carried out. The inclusion criteria were implementing head and neck medical images (computed tomography (CT), positron emission tomography (PET), MRI, Planar scans, and panoramic X-ray) of human subjects with segmentation, object detection, and classification deep learning models for head and neck cancers. The risk of bias was rated with the quality assessment of diagnostic accuracy studies (QUADAS-2) tool. For the meta-analysis diagnostic odds ratio (DOR) was calculated. Deeks' funnel plot was used to assess publication bias. MIDAS and Metandi packages were used to analyze diagnostic test accuracy in STATA. RESULTS: From 1967 studies, 32 were found eligible after the search and screening procedures. According to the QUADAS-2 tool, 7 included studies had a low risk of bias for all domains. According to the results of all included studies, the accuracy varied from 82.6 to 100%. Additionally, specificity ranged from 66.6 to 90.1%, sensitivity from 74 to 99.68%. Fourteen studies that provided sufficient data were included for meta-analysis. The pooled sensitivity was 90% (95% CI 0.820.94), and the pooled specificity was 92% (CI 95% 0.87-0.96). The DORs were 103 (27-251). Publication bias was not detected based on the p-value of 0.75 in the meta-analysis. CONCLUSION: With a head and neck screening deep learning model, detectable screening processes can be enhanced with high specificity and sensitivity.


Assuntos
Aprendizado Profundo , Neoplasias de Cabeça e Pescoço , Humanos , Sensibilidade e Especificidade , Neoplasias de Cabeça e Pescoço/diagnóstico por imagem , Imageamento por Ressonância Magnética/métodos , Tomografia por Emissão de Pósitrons/métodos
13.
Int Endod J ; 57(3): 305-314, 2024 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-38117284

RESUMO

AIM: This study aimed to evaluate and compare the validity and reliability of responses provided by GPT-3.5, Google Bard, and Bing to frequently asked questions (FAQs) in the field of endodontics. METHODOLOGY: FAQs were formulated by expert endodontists (n = 10) and collected through GPT-3.5 queries (n = 10), with every question posed to each chatbot three times. Responses (N = 180) were independently evaluated by two board-certified endodontists using a modified Global Quality Score (GQS) on a 5-point Likert scale (5: strongly agree; 4: agree; 3: neutral; 2: disagree; 1: strongly disagree). Disagreements on scoring were resolved through evidence-based discussions. The validity of responses was analysed by categorizing scores into valid or invalid at two thresholds: The low threshold was set at score ≥4 for all three responses whilst the high threshold was set at score 5 for all three responses. Fisher's exact test was conducted to compare the validity of responses between chatbots. Cronbach's alpha was calculated to assess the reliability by assessing the consistency of repeated responses for each chatbot. RESULTS: All three chatbots provided answers to all questions. Using the low-threshold validity test (GPT-3.5: 95%; Google Bard: 85%; Bing: 75%), there was no significant difference between the platforms (p > .05). When using the high-threshold validity test, the chatbot scores were substantially lower (GPT-3.5: 60%; Google Bard: 15%; Bing: 15%). The validity of GPT-3.5 responses was significantly higher than Google Bard and Bing (p = .008). All three chatbots achieved an acceptable level of reliability (Cronbach's alpha >0.7). CONCLUSIONS: GPT-3.5 provided more credible information on topics related to endodontics compared to Google Bard and Bing.


Assuntos
Inteligência Artificial , Endodontia , Reprodutibilidade dos Testes , Software , Fonte de Informação
14.
Sci Rep ; 13(1): 13755, 2023 08 23.
Artigo em Inglês | MEDLINE | ID: mdl-37612309

RESUMO

Digital images allow for the objective evaluation of facial appearance and abnormalities as well as treatment outcomes and stability. With the advancement of technology, manual clinical measurements can be replaced with fully automatic photographic assessments. However, obtaining millimetric measurements on photographs does not provide clinicians with their actual value due to different image magnification ratios. A deep learning tool was developed to estimate linear measurements on images with unknown magnification using the iris diameter. A framework was designed to segment the eyes' iris and calculate the horizontal visible iris diameter (HVID) in pixels. A constant value of 12.2 mm was assigned as the HVID value in all the photographs. A vertical and a horizontal distance were measured in pixels on photographs of 94 subjects and were estimated in millimeters by calculating the magnification ratio using HVID. Manual measurement of the distances was conducted on the subjects and the actual and estimated amounts were compared using Bland-Altman analysis. The obtained error was calculated as mean absolute percentage error (MAPE) of 2.9% and 4.3% in horizontal and vertical measurements. Our study shows that due to the consistent size and narrow range of HVID values, the iris diameter can be used as a reliable scale to calibrate the magnification of the images to obtain precise measurements in further research.


Assuntos
Aprendizado Profundo , Gênero Iris , Humanos , Irã (Geográfico) , Face , Fácies , Iris
15.
J Dent ; 135: 104593, 2023 08.
Artigo em Inglês | MEDLINE | ID: mdl-37355089

RESUMO

OBJECTIVE: Artificial Intelligence (AI) refers to the ability of machines to perform cognitive and intellectual human tasks. In dentistry, AI offers the potential to enhance diagnostic accuracy, improve patient outcomes and streamline workflows. The present study provides a framework and a checklist to evaluate AI applications in dentistry from this perspective. METHODS: Lending from existing guidance documents, an initial draft of the checklist and an explanatory paper were derived and discussed among the groups members. RESULTS: The checklist was consented to in an anonymous voting process by 29 Topic Group Dental Diagnostics and Digital Dentistry, ITU/WHO Focus Group AI on Health's members. Overall, 11 principles were identified (diversity, transparency, wellness, privacy protection, solidarity, equity, prudence, law and governance, sustainable development, accountability, and responsibility, respect of autonomy, decision-making). CONCLUSIONS: Providers, patients, researchers, industry, and other stakeholders should consider these principles when developing, implementing, or receiving AI applications in dentistry. CLINICAL SIGNIFICANCE: While AI has become increasingly commonplace in dentistry, there are ethical concerns around its usage, and users (providers, patients, and other stakeholders), as well as the industry should consider these when developing, implementing, or receiving AI applications based on comprehensive framework to address the associated ethical challenges.


Assuntos
Inteligência Artificial , Lista de Checagem , Humanos , Grupos Focais , Privacidade , Odontologia
16.
J Dent ; 135: 104556, 2023 08.
Artigo em Inglês | MEDLINE | ID: mdl-37209769

RESUMO

OBJECTIVE: Federated Learning (FL) enables collaborative training of artificial intelligence (AI) models from multiple data sources without directly sharing data. Due to the large amount of sensitive data in dentistry, FL may be particularly relevant for oral and dental research and applications. This study, for the first time, employed FL for a dental task, automated tooth segmentation on panoramic radiographs. METHODS: We employed a dataset of 4,177 panoramic radiographs collected from nine different centers (n = 143 to n = 1881 per center) across the globe and used FL to train a machine learning model for tooth segmentation. FL performance was compared against Local Learning (LL), i.e., training models on isolated data from each center (assuming data sharing not to be an option). Further, the performance gap to Central Learning (CL), i.e., training on centrally pooled data (based on data sharing agreements) was quantified. Generalizability of models was evaluated on a pooled test dataset from all centers. RESULTS: For 8 out of 9 centers, FL outperformed LL with statistical significance (p<0.05); only the center providing the largest amount of data FL did not have such an advantage. For generalizability, FL outperformed LL across all centers. CL surpassed both FL and LL for performance and generalizability. CONCLUSION: If data pooling (for CL) is not feasible, FL is shown to be a useful alternative to train performant and, more importantly, generalizable deep learning models in dentistry, where data protection barriers are high. CLINICAL SIGNIFICANCE: This study proves the validity and utility of FL in the field of dentistry, which encourages researchers to adopt this method to improve the generalizability of dental AI models and ease their transition to the clinical environment.


Assuntos
Inteligência Artificial , Aprendizado Profundo , Humanos , Radiografia Panorâmica , Pesquisadores
17.
Diagnostics (Basel) ; 13(5)2023 Mar 06.
Artigo em Inglês | MEDLINE | ID: mdl-36900140

RESUMO

Using super-resolution (SR) algorithms, an image with a low resolution can be converted into a high-quality image. Our objective was to compare deep learning-based SR models to a conventional approach for improving the resolution of dental panoramic radiographs. A total of 888 dental panoramic radiographs were obtained. Our study involved five state-of-the-art deep learning-based SR approaches, including SR convolutional neural networks (SRCNN), SR generative adversarial network (SRGAN), U-Net, Swin for image restoration (SwinIr), and local texture estimator (LTE). Their results were compared with one another and with conventional bicubic interpolation. The performance of each model was evaluated using the metrics of mean squared error (MSE), peak signal-to-noise ratio (PNSR), structural similarity index (SSIM), and mean opinion score by four experts (MOS). Among all the models evaluated, the LTE model presented the highest performance, with MSE, SSIM, PSNR, and MOS results of 7.42 ± 0.44, 39.74 ± 0.17, 0.919 ± 0.003, and 3.59 ± 0.54, respectively. Additionally, compared with low-resolution images, the output of all the used approaches showed significant improvements in MOS evaluation. A significant enhancement in the quality of panoramic radiographs can be achieved by SR. The LTE model outperformed the other models.

18.
J Dent ; 130: 104430, 2023 03.
Artigo em Inglês | MEDLINE | ID: mdl-36682721

RESUMO

OBJECTIVES: Despite deep learning's wide adoption in dental artificial intelligence (AI) research, researchers from other dental fields and, more so, dental professionals may find it challenging to understand and interpret deep learning studies, their employed methods, and outcomes. The objective of this primer is to explain the basic concept of deep learning. It will lay out the commonly used terms, and describe different deep learning approaches, their methods, and outcomes. METHODS: Our research is based on the latest review studies, medical primers, as well as the state-of-the-art research on AI and deep learning, which have been gathered in the current study. RESULTS: In this study, a basic understanding of deep learning models and various approaches to deep learning is presented. An overview of data management strategies for deep learning projects is presented, including data collection, data curation, data annotation, and data preprocessing. Additionally, we provided a step-by-step guide for completing a real-world project. CONCLUSION: Researchers and clinicians can benefit from this study by gaining insight into deep learning. It can be used to critically appraise existing work or plan new deep learning projects. CLINICAL SIGNIFICANCE: This study may be useful to dental researchers and professionals who are assessing and appraising deep learning studies within the field of dentistry.


Assuntos
Inteligência Artificial , Aprendizado Profundo , Humanos , Odontólogos
19.
J Endod ; 49(3): 248-261.e3, 2023 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-36563779

RESUMO

INTRODUCTION: The aim of this systematic review and meta-analysis was to investigate the overall accuracy of deep learning models in detecting periapical (PA) radiolucent lesions in dental radiographs, when compared to expert clinicians. METHODS: Electronic databases of Medline (via PubMed), Embase (via Ovid), Scopus, Google Scholar, and arXiv were searched. Quality of eligible studies was assessed by using Quality Assessment and Diagnostic Accuracy Tool-2. Quantitative analyses were conducted using hierarchical logistic regression for meta-analyses on diagnostic accuracy. Subgroup analyses on different image modalities (PA radiographs, panoramic radiographs, and cone beam computed tomographic images) and on different deep learning tasks (classification, segmentation, object detection) were conducted. Certainty of evidence was assessed by using Grading of Recommendations Assessment, Development, and Evaluation system. RESULTS: A total of 932 studies were screened. Eighteen studies were included in the systematic review, out of which 6 studies were selected for quantitative analyses. Six studies had low risk of bias. Twelve studies had risk of bias. Pooled sensitivity, specificity, positive likelihood ratio, negative likelihood ratio, and diagnostic odds ratio of included studies (all image modalities; all tasks) were 0.925 (95% confidence interval [CI], 0.862-0.960), 0.852 (95% CI, 0.810-0.885), 6.261 (95% CI, 4.717-8.311), 0.087 (95% CI, 0.045-0.168), and 71.692 (95% CI, 29.957-171.565), respectively. No publication bias was detected (Egger's test, P = .82). Grading of Recommendations Assessment, Development and Evaluationshowed a "high" certainty of evidence for the studies included in the meta-analyses. CONCLUSION: Compared to expert clinicians, deep learning showed highly accurate results in detecting PA radiolucent lesions in dental radiographs. Most studies had risk of bias. There was a lack of prospective studies.


Assuntos
Aprendizado Profundo , Tomografia Computadorizada de Feixe Cônico/métodos , Radiografia Panorâmica , Testes Diagnósticos de Rotina , Sensibilidade e Especificidade
20.
Dent Res J (Isfahan) ; 20: 116, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-38169618

RESUMO

Background: Dentists begin the diagnosis by identifying and enumerating teeth. Panoramic radiographs are widely used for tooth identification due to their large field of view and low exposure dose. The automatic numbering of teeth in panoramic radiographs can assist clinicians in avoiding errors. Deep learning has emerged as a promising tool for automating tasks. Our goal is to evaluate the accuracy of a two-step deep learning method for tooth identification and enumeration in panoramic radiographs. Materials and Methods: In this retrospective observational study, 1007 panoramic radiographs were labeled by three experienced dentists. It involved drawing bounding boxes in two distinct ways: one for teeth and one for quadrants. All images were preprocessed using the contrast-limited adaptive histogram equalization method. First, panoramic images were allocated to a quadrant detection model, and the outputs of this model were provided to the tooth numbering models. A faster region-based convolutional neural network model was used in each step. Results: Average precision (AP) was calculated in different intersection-over-union thresholds. The AP50 of quadrant detection and tooth enumeration was 100% and 95%, respectively. Conclusion: We have obtained promising results with a high level of AP using our two-step deep learning framework for automatic tooth enumeration on panoramic radiographs. Further research should be conducted on diverse datasets and real-life situations.

SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA
...