Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 85
Filtrar
1.
Imaging Sci Dent ; 54(3): 276-282, 2024 Sep.
Artigo em Inglês | MEDLINE | ID: mdl-39371311

RESUMO

Purpose: This study aimed to assess the performance of 2-dimensional (2D) imaging with microscopy coils in delineating teeth and periodontal tissues compared with conventional 3-dimensional (3D) imaging on a 3 T magnetic resonance imaging (MRI) unit. Materials and Methods: Twelve healthy participants (4 men and 8 women; mean age: 25.6 years; range: 20-52 years) with no dental symptoms were included. The left mandibular first molars and surrounding periodontal tissues were examined using the following 2 sequences: 2D proton density-weighted (PDw) images and 3D enhanced T1 high-resolution isotropic volume excitation (eTHRIVE) images. Two-dimensional MRI images were taken using a 3 T MRI unit and a 47 mm microscopy coil, while 3D MRI imaging used a 3 T MRI unit and head-neck coil. Oral radiologists assessed dental and periodontal structures using a 4-point Likert scale. Inter- and intra-observer agreement was determined using the weighted kappa coefficient. The Wilcoxon signed-rank test was used to compare 2D-PDw and 3D-eTHRIVE images. Results: Qualitative analysis showed significantly better visualization scores for 2D-PDw imaging than for 3D-eTHRIVE imaging (Wilcoxon signed-rank test). 2D-PDw images provided improved visibility of the tooth, root dental pulp, periodontal ligament, lamina dura, coronal dental pulp, gingiva, and nutrient tract. Inter-observer reliability ranged from moderate agreement to almost perfect agreement, and intra-observer agreement was in a similar range. Conclusion: Two-dimensional-PDw images acquired using a 3 T MRI unit and microscopy coil effectively visualized nearly all aspects of teeth and periodontal tissues.

2.
Artigo em Inglês | MEDLINE | ID: mdl-39222427

RESUMO

OBJECTIVES: The purpose of this study was to generate radiographs including dentigerous cysts by applying the latest generative adversarial network (GAN; StyleGAN3) to panoramic radiography. METHODS: A total of 459 cystic lesions were selected, and 409 images were randomly assigned as training data and 50 images as test data. StyleGAN3 training was performed for 500 000 images. Fifty generated images were objectively evaluated by comparing them with 50 real images according to four metrics: Fréchet inception distance (FID), kernel inception distance (KID), precision and recall, and inception score (IS). A subjective evaluation of the generated images was performed by three specialists who compared them with the real images in a visual Turing test. RESULTS: The results of the metrics were as follows: FID, 199.28; KID, 0.14; precision, 0.0047; recall, 0.00; and IS, 2.48. The overall results of the visual Turing test were 82.3%. No significant difference was found in the human scoring of root resorption. CONCLUSIONS: The images generated by StyleGAN3 were of such high quality that specialists could not distinguish them from the real images.

3.
Odontology ; 2024 Aug 29.
Artigo em Inglês | MEDLINE | ID: mdl-39198339

RESUMO

The purpose of this study is to develop two-step deep learning models that can automatically detect implant regions on panoramic radiographs and identify several types of implants. A total of 1,574 panoramic radiographs containing 3675 implants were included. The implant manufacturers were Kyocera, Dentsply Sirona, Straumann, and Nobel Biocare. Model A was created to detect oral implants and identify the manufacturers using You Only Look Once (YOLO) v7. After preparing the image patches that cropped the implant regions detected by model A, model B was created to identify the implant types per manufacturer using EfficientNet. Model A achieved very high performance, with recall of 1.000, precision of 0.979, and F1 score of 0.989. It also had accuracy, recall, precision, and F1 score of 0.98 or higher for the classification of the manufacturers. Model B had high classification metrics above 0.92, exception for Nobel's class 2 (Parallel). In this study, two-step deep learning models were built to detect implant regions, identify four manufacturers, and identify implant types per manufacturer.

4.
Artigo em Inglês | MEDLINE | ID: mdl-39067043

RESUMO

OBJECTIVES: This study aimed to clarify the performance of magnetic resonance imaging (MRI)-based deep learning classification models in diagnosing temporomandibular joint osteoarthritis (TMJ-OA) and to compare the developed diagnostic assistance with human observers. METHODS: The subjects were 118 patients who underwent MRI for examination of TMJ disorders. One hundred condyles with TMJ-OA and 100 condyles without TMJ-OA were enrolled. Deep learning was performed with four networks (ResNet18, EfficientNet b4, Inception v3, and GoogLeNet) using five-fold cross validation. Receiver operating characteristics (ROC) curves were drawn for each model and diagnostic metrics were determined. The performances of the four network models were compared using Kruskal-Wallis tests and post-hoc Scheffe tests, and ROCs between the best model and human were compared using chi-square tests, with p < 0.05 considered significant. RESULTS: ResNet18 had areas under the curves (AUCs) of 0.91-0.93 and accuracy of 0.85-0.88, which were the highest among the four networks. There were significant differences in AUC and accuracy between ResNet and GoogLeNet (p = 0.0264 and p = 0.0418, respectively). The kappa values of the models were large, 0.95 for ResNet and 0.93 for EfficientNet. The experts achieved similar AUC and accuracy values ​​to the ResNet metrics, 0.94 and 0.85, and 0.84 and 0.84, respectively, but with a lower kappa of 0.67. Those of the dental residents showed lower values. There were significant differences in AUCs between ResNet and residents (p < 0.0001) and between experts and residents (p < 0.0001). CONCLUSIONS: Using a deep learning model, high performance was confirmed for MRI diagnosis of TMJ-OA.

5.
Oral Radiol ; 40(4): 538-545, 2024 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-38990220

RESUMO

OBJECTIVE: The present study aimed to assess the consistencies and performances of deep learning (DL) models in the diagnosis of condylar osteoarthritis (OA) among patients with dentofacial deformities using panoramic temporomandibular joint (TMJ) projection images. METHODS: A total of 68 TMJs with or without condylar OA in dentofacial deformity patients were tested to verify the consistencies and performances of DL models created using 252 TMJs with or without OA in TMJ disorder and dentofacial deformity patients; these models were used to diagnose OA on conventional panoramic (Con-Pa) images and open (Open-TMJ) and closed (Closed-TMJ) mouth TMJ projection images. The GoogLeNet and VGG-16 networks were used to create the DL models. For comparison, two dental residents with < 1 year of experience interpreting radiographs evaluated the same condyle data that had been used to test the DL models. RESULTS: On Open-TMJ images, the DL models showed moderate to very good consistency, whereas the residents' demonstrated fair consistency on all images. The areas under the curve (AUCs) of both DL models on Con-Pa (0.84 for GoogLeNet and 0.75 for VGG-16) and Open-TMJ images (0.89 for both models) were significantly higher than the residents' AUCs (p < 0.01). The AUCs of the DL models on Open-TMJ images (0.89 for both models) were higher than the AUCs on Closed-TMJ images (0.72 for both models). CONCLUSIONS: The DL models created in this study could help residents to interpret Con-Pa and Open-TMJ images in the diagnosis of condylar OA.


Assuntos
Aprendizado Profundo , Deformidades Dentofaciais , Côndilo Mandibular , Osteoartrite , Radiografia Panorâmica , Humanos , Osteoartrite/diagnóstico por imagem , Feminino , Masculino , Côndilo Mandibular/diagnóstico por imagem , Adulto , Pessoa de Meia-Idade , Deformidades Dentofaciais/diagnóstico por imagem , Transtornos da Articulação Temporomandibular/diagnóstico por imagem , Adolescente , Articulação Temporomandibular/diagnóstico por imagem , Idoso , Adulto Jovem
6.
Oral Radiol ; 40(4): 508-519, 2024 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-38941003

RESUMO

OBJECTIVES: The objective of this study was to enhance the visibility of soft tissues on cone-beam computed tomography (CBCT) using a CycleGAN network trained on CT images. METHODS: Training and evaluation of the CycleGAN were conducted using CT and CBCT images collected from Aichi Gakuin University (α facility) and Osaka Dental University (ß facility). Synthesized images (sCBCT) output by the CycleGAN network were evaluated by comparing them with the original images (oCBCT) and CT images, and assessments were made using histogram analysis and human scoring of soft-tissue anatomical structures and cystic lesions. RESULTS: The histogram analysis showed that on sCBCT, soft-tissue anatomical structures showed significant shifts in voxel intensity toward values resembling those on CT, with the mean values for all structures approaching those of CT and the specialists' visibility scores being significantly increased. However, improvement in the visibility of cystic lesions was limited. CONCLUSIONS: Image synthesis using CycleGAN significantly improved the visibility of soft tissue on CBCT, with this improvement being particularly notable from the submandibular region to the floor of the mouth. Although the effect on the visibility of cystic lesions was limited, there is potential for further improvement through refinement of the training method.


Assuntos
Tomografia Computadorizada de Feixe Cônico , Humanos , Inteligência Artificial , Interpretação de Imagem Radiográfica Assistida por Computador , Redes Neurais de Computação , Feminino , Masculino
7.
Odontology ; 112(4): 1343-1352, 2024 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-38607582

RESUMO

The objectives of this study were to create a mutual conversion system between contrast-enhanced computed tomography (CECT) and non-CECT images using a cycle generative adversarial network (cycleGAN) for the internal jugular region. Image patches were cropped from CT images in 25 patients who underwent both CECT and non-CECT imaging. Using a cycleGAN, synthetic CECT and non-CECT images were generated from original non-CECT and CECT images, respectively. The peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) were calculated. Visual Turing tests were used to determine whether oral and maxillofacial radiologists could tell the difference between synthetic versus original images, and receiver operating characteristic (ROC) analyses were used to assess the radiologists' performances in discriminating lymph nodes from blood vessels. The PSNR of non-CECT images was higher than that of CECT images, while the SSIM was higher in CECT images. The Visual Turing test showed a higher perceptual quality in CECT images. The area under the ROC curve showed almost perfect performances in synthetic as well as original CECT images. In conclusion, synthetic CECT images created by cycleGAN appeared to have the potential to provide effective information in patients who could not receive contrast enhancement.


Assuntos
Meios de Contraste , Tomografia Computadorizada por Raios X , Humanos , Tomografia Computadorizada por Raios X/métodos , Masculino , Feminino , Linfonodos/diagnóstico por imagem , Razão Sinal-Ruído , Pessoa de Meia-Idade , Idoso , Redes Neurais de Computação , Curva ROC , Adulto , Veias Jugulares/diagnóstico por imagem
8.
Imaging Sci Dent ; 54(1): 33-41, 2024 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-38571775

RESUMO

Purpose: The aims of this study were to create a deep learning model to distinguish between nasopalatine duct cysts (NDCs), radicular cysts, and no-lesions (normal) in the midline region of the anterior maxilla on panoramic radiographs and to compare its performance with that of dental residents. Materials and Methods: One hundred patients with a confirmed diagnosis of NDC (53 men, 47 women; average age, 44.6±16.5 years), 100 with radicular cysts (49 men, 51 women; average age, 47.5±16.4 years), and 100 with normal groups (56 men, 44 women; average age, 34.4±14.6 years) were enrolled in this study. Cases were randomly assigned to the training datasets (80%) and the test dataset (20%). Then, 20% of the training data were randomly assigned as validation data. A learning model was created using a customized DetectNet built in Digits version 5.0 (NVIDIA, Santa Clara, USA). The performance of the deep learning system was assessed and compared with that of two dental residents. Results: The performance of the deep learning system was superior to that of the dental residents except for the recall of radicular cysts. The areas under the curve (AUCs) for NDCs and radicular cysts in the deep learning system were significantly higher than those of the dental residents. The results for the dental residents revealed a significant difference in AUC between NDCs and normal groups. Conclusion: This study showed superior performance in detecting NDCs and radicular cysts and in distinguishing between these lesions and normal groups.

9.
Imaging Sci Dent ; 54(1): 25-31, 2024 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-38571781

RESUMO

Purpose: The purpose of this study was to clarify the panoramic image differences of cleft alveolus patients with or without a cleft palate, with emphases on the visibility of the line formed by the junction between the nasal septum and nasal floor (the upper line) and the appearances of the maxillary lateral incisor. Materials and Methods: Panoramic radiographs of 238 patients with cleft alveolus were analyzed for the visibility of the upper line, including clear, obscure or invisible, and the appearances of the maxillary lateral incisor, regarding congenital absence, incomplete growth, delayed eruption and medial inclination. Differences in the distribution ratio of these visibility and appearances were verified between the patients with and without a cleft palate using the chi-square test. Results: There was a significant difference in the visibility distribution of the upper line between the patients with and without a cleft palate (p<0.05). In most of the patients with a cleft palate, the upper line was not observed. In the unilateral cleft alveolus patients, the medial inclination of the maxillary lateral incisor was more frequently observed in patients with a cleft palate than in patients without a cleft palate. Conclusion: Two differences were identified in panoramic appearances. The first was the disappearance (invisible appearance) of the upper line in patients with a cleft palate, and the second was a change in the medial inclination on the affected side maxillary lateral incisor in unilateral cleft alveolus patients with a cleft palate.

10.
J Endod ; 50(5): 627-636, 2024 May.
Artigo em Inglês | MEDLINE | ID: mdl-38336338

RESUMO

INTRODUCTION: The purposes of this study were to evaluate the effect of the combined use of object detection for the classification of the C-shaped canal anatomy of the mandibular second molar in panoramic radiographs and to perform an external validation on a multicenter dataset. METHODS: The panoramic radiographs of 805 patients were collected from 4 institutes across two countries. The CBCT data of the same patients were used as "Ground-truth". Five datasets were generated: one for training and validation, and 4 as external validation datasets. Workflow 1 used manual cropping to prepare the image patches of mandibular second molars, and then classification was performed using EfficientNet. Workflow 2 used two combined methods with a preceding object detection (YOLOv7) performed for automated image patch formation, followed by classification using EfficientNet. Workflow 3 directly classified the root canal anatomy from the panoramic radiographs using the YOLOv7 prediction outcomes. The classification performance of the 3 workflows was evaluated and compared across 4 external validation datasets. RESULTS: For Workflows 1, 2, and 3, the area under the receiver operating characteristic curve (AUC) values were 0.863, 0.861, and 0.876, respectively, for the AGU dataset; 0.935, 0.945, and 0.863, respectively, for the ASU dataset; 0.854, 0.857, and 0.849, respectively, for the ODU dataset; and 0.821, 0.797, and 0.831, respectively, for the ODU low-resolution dataset. No significant differences existed between the AUC values of Workflows 1, 2, and 3 across the 4 datasets. CONCLUSIONS: The deep learning systems of the 3 workflows achieved significant accuracy in predicting the C-shaped canal in mandibular second molars across all test datasets.


Assuntos
Cavidade Pulpar , Mandíbula , Dente Molar , Radiografia Panorâmica , Humanos , Dente Molar/diagnóstico por imagem , Dente Molar/anatomia & histologia , Mandíbula/diagnóstico por imagem , Mandíbula/anatomia & histologia , Cavidade Pulpar/diagnóstico por imagem , Cavidade Pulpar/anatomia & histologia , Feminino , Masculino , Tomografia Computadorizada de Feixe Cônico/métodos , Adulto
11.
Cancers (Basel) ; 16(2)2024 Jan 08.
Artigo em Inglês | MEDLINE | ID: mdl-38254765

RESUMO

Ultrasonography is the preferred modality for detailed evaluation of enlarged lymph nodes (LNs) identified on computed tomography and/or magnetic resonance imaging, owing to its high spatial resolution. However, the diagnostic performance of ultrasonography depends on the examiner's expertise. To support the ultrasonographic diagnosis, we developed YOLOv7-based deep learning models for metastatic LN detection on ultrasonography and compared their detection performance with that of highly experienced radiologists and less experienced residents. We enrolled 462 B- and D-mode ultrasound images of 261 metastatic and 279 non-metastatic histopathologically confirmed LNs from 126 patients with head and neck squamous cell carcinoma. The YOLOv7-based B- and D-mode models were optimized using B- and D-mode training and validation images and their detection performance for metastatic LNs was evaluated using B- and D-mode testing images, respectively. The D-mode model's performance was comparable to that of radiologists and superior to that of residents' reading of D-mode images, whereas the B-mode model's performance was higher than that of residents but lower than that of radiologists on B-mode images. Thus, YOLOv7-based B- and D-mode models can assist less experienced residents in ultrasonographic diagnoses. The D-mode model could raise the diagnostic performance of residents to the same level as experienced radiologists.

12.
Imaging Sci Dent ; 53(1): 27-34, 2023 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-37006785

RESUMO

Purpose: The aim of this study was to clarify the influence of training with a different kind of lesion on the performance of a target model. Materials and Methods: A total of 310 patients (211 men, 99 women; average age, 47.9±16.1 years) were selected and their panoramic images were used in this study. We created a source model using panoramic radiographs including mandibular radiolucent cyst-like lesions (radicular cyst, dentigerous cyst, odontogenic keratocyst, and ameloblastoma). The model was simulatively transferred and trained on images of Stafne's bone cavity. A learning model was created using a customized DetectNet built in the Digits version 5.0 (NVIDIA, Santa Clara, CA). Two machines (Machines A and B) with identical specifications were used to simulate transfer learning. A source model was created from the data consisting of ameloblastoma, odontogenic keratocyst, dentigerous cyst, and radicular cyst in Machine A. Thereafter, it was transferred to Machine B and trained on additional data of Stafne's bone cavity to create target models. To investigate the effect of the number of cases, we created several target models with different numbers of Stafne's bone cavity cases. Results: When the Stafne's bone cavity data were added to the training, both the detection and classification performances for this pathology improved. Even for lesions other than Stafne's bone cavity, the detection sensitivities tended to increase with the increase in the number of Stafne's bone cavities. Conclusion: This study showed that using different lesions for transfer learning improves the performance of the model.

13.
Oral Radiol ; 39(3): 467-474, 2023 07.
Artigo em Inglês | MEDLINE | ID: mdl-36166134

RESUMO

OBJECTIVES: To clarify the performance of transfer learning with a small number of Waters' images at institution B in diagnosing maxillary sinusitis, based on a source model trained with a large number of panoramic radiographs at institution A. METHODS: The source model was created by a 200-epoch training process with 800 training and 60 validation datasets of panoramic radiographs at institution A using VGG-16. One hundred and eighty Waters' and 180 panoramic image patches with or without maxillary sinusitis at institution B were enrolled in this study, and were arbitrarily assigned to 120 training, 20 validation, and 40 test datasets, respectively. Transfer learning of 200 epochs was performed using the training and validation datasets of Waters' images based on the source model, and the target model was obtained. The test Waters' images were applied to the source and target models, and the performance of each model was evaluated. Transfer learning with panoramic radiographs and evaluation by two radiologists were undertaken and compared. The evaluation was based on the area of receiver-operating characteristic curves (AUC). RESULTS: When using Waters' images as the test dataset, the AUCs of the source model, target model, and radiologists were 0.780, 0.830, and 0.806, respectively. There were no significant differences between these models and the radiologists, whereas the target model performed better than the source model. For panoramic radiographs, AUCs were 0.863, 0.863, and 0.808, respectively, with no significant differences. CONCLUSIONS: This study performed transfer learning using a small number of Waters' images, based on a source model created solely from panoramic radiographs, resulting in a performance improvement to 0.830 in diagnosing maxillary sinusitis, which was equivalent to that of radiologists. Transfer learning is considered a useful method to improve diagnostic performance.


Assuntos
Aprendizado Profundo , Sinusite Maxilar , Humanos , Sinusite Maxilar/diagnóstico por imagem , Radiografia Panorâmica , Radiografia , Radiologistas
14.
Dentomaxillofac Radiol ; 52(8): 20210436, 2023 Nov.
Artigo em Inglês | MEDLINE | ID: mdl-35076259

RESUMO

OBJECTIVES: The purpose of this study was to evaluate the difference in performance of deep-learning (DL) models with respect to the image classes and amount of training data to create an effective DL model for detecting both unilateral cleft alveoli (UCAs) and bilateral cleft alveoli (BCAs) on panoramic radiographs. METHODS: Model U was created using UCA and normal images, and Model B was created using BCA and normal images. Models C1 and C2 were created using the combined data of UCA, BCA, and normal images. The same number of CAs was used for training Models U, B, and C1, whereas Model C2 was created with a larger amount of data. The performance of all four models was evaluated with the same test data and compared with those of two human observers. RESULTS: The recall values were 0.60, 0.73, 0.80, and 0.88 for Models A, B, C1, and C2, respectively. The results of Model C2 were highest in precision and F-measure (0.98 and 0.92) and almost the same as those of human observers. Significant differences were found in the ratios of detected to undetected CAs of Models U and C1 (p = 0.01), Models U and C2 (p < 0.001), and Models B and C2 (p = 0.036). CONCLUSIONS: The DL models trained using both UCA and BCA data (Models C1 and C2) achieved high detection performance. Moreover, the performance of a DL model may depend on the amount of training data.


Assuntos
Aprendizado Profundo , Humanos , Radiografia Panorâmica
16.
Oral Radiol ; 39(2): 349-354, 2023 04.
Artigo em Inglês | MEDLINE | ID: mdl-35984588

RESUMO

OBJECTIVES: The aim of the present study was to create effective deep learning-based models for diagnosing the presence or absence of cleft palate (CP) in patients with unilateral or bilateral cleft alveolus (CA) on panoramic radiographs. METHODS: The panoramic images of 491 patients who had unilateral or bilateral cleft alveolus were used to create two models. Model A, which detects the upper incisor area on panoramic radiographs and classifies the areas into the presence or absence of CP, was created using both object detection and classification functions of DetectNet. Using the same data for developing Model A, Model B, which directly classifies the presence or absence of CP on panoramic radiographs, was created using classification function of VGG-16. The performances of both models were evaluated with the same test data and compared with those of two radiologists. RESULTS: The recall, precision, and F-measure were all 1.00 in Model A. The area under the receiver operating characteristic curve (AUC) values were 0.95, 0.93, 0.70, and 0.63 for Model A, Model B, and the radiologists, respectively. The AUCs of the models were significantly higher than those of the radiologists. CONCLUSIONS: The deep learning-based models developed in the present study have potential for use in supporting observer interpretations of the presence of cleft palate on panoramic radiographs.


Assuntos
Fissura Palatina , Aprendizado Profundo , Humanos , Fissura Palatina/diagnóstico por imagem , Radiografia Panorâmica , Incisivo
17.
Odontology ; 111(1): 228-236, 2023 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-35951139

RESUMO

This study aimed to determine the association between the progressive contraction of the posterior pharyngeal wall and dysphagia in postoperative patients with tongue cancer. A videofluoroscopic swallowing study (VFSS) was performed in 34 patients after tongue cancer surgery. Images were analyzed using a two-dimensional video measurement software. Cases in which the processes on the posterior pharyngeal wall moved downward from the 2nd to 4th vertebral regions were defined as "normal type", other cases were defined as "abnormal type". Twenty-four patients showed normal movement of the posterior pharyngeal wall, whereas 10 patients showed the abnormal type. The results showed that there was a significant difference in dysphagia scores between the postoperative swallowing type and swallowing dysfunction score. This implies that dysphagia is related to the movement of the posterior pharyngeal wall after tongue cancer surgery. Furthermore, the extent of resection and stage were significantly different between the normal and abnormal groups in the posterior pharyngeal wall movement. There was also a significant difference between the two groups in terms of the following: whether the tongue base was included in the excision range (p < 0.01), whether neck dissection was performed (p < 0.01), or whether reconstruction was not performed (p < 0.01). VFSS results showed that posterior pharyngeal wall movement was altered after surgery in patients with tongue cancer who had severe dysphagia.


Assuntos
Transtornos de Deglutição , Deglutição , Fluoroscopia , Neoplasias da Língua , Humanos , Transtornos de Deglutição/diagnóstico por imagem , Transtornos de Deglutição/etiologia , Faringe/diagnóstico por imagem , Língua , Neoplasias da Língua/cirurgia
18.
Sci Rep ; 12(1): 18754, 2022 11 05.
Artigo em Inglês | MEDLINE | ID: mdl-36335226

RESUMO

Although videofluorography (VFG) is an effective tool for evaluating swallowing functions, its accurate evaluation requires considerable time and effort. This study aimed to create a deep learning model for automated bolus segmentation on VFG images of patients with healthy swallowing and dysphagia using the artificial intelligence deep learning segmentation method, and to assess the performance of the method. VFG images of 72 swallowing of 12 patients were continuously converted into 15 static images per second. In total, 3910 images were arbitrarily assigned to the training, validation, test 1, and test 2 datasets. In the training and validation datasets, images of colored bolus areas were prepared, along with original images. Using a U-Net neural network, a trained model was created after 500 epochs of training. The test datasets were applied to the trained model, and the performances of automatic segmentation (Jaccard index, Sørensen-Dice coefficient, and sensitivity) were calculated. All performance values for the segmentation of the test 1 and 2 datasets were high, exceeding 0.9. Using an artificial intelligence deep learning segmentation method, we automatically segmented the bolus areas on VFG images; our method exhibited high performance. This model also allowed assessment of aspiration and laryngeal invasion.


Assuntos
Aprendizado Profundo , Humanos , Processamento de Imagem Assistida por Computador/métodos , Deglutição , Inteligência Artificial , Redes Neurais de Computação
19.
Artigo em Inglês | MEDLINE | ID: mdl-36229373

RESUMO

OBJECTIVE: The aim of this study was to create and assess a deep learning model using segmentation and transfer learning methods to visualize the proximity of the mandibular canal to an impacted third molar on panoramic radiographs. STUDY DESIGN: The panoramic radiographs containing the mandibular canal and impacted third molar were collected from 2 hospitals (Hospitals A and B). A total of 3200 areas were used for creating and evaluating learning models. A source model was created using the data from Hospital A, simulatively transferred to Hospital B, and trained using various amounts of data from Hospital B to create target models. The same data were then applied to the target models to calculate the Dice coefficient, Jaccard index, and sensitivity. RESULTS: The performance of target models trained using 200 or more data sets was equivalent to that of the source model tested using data obtained from the same hospital (Hospital A). CONCLUSIONS: Sufficiently qualified models could delineate the mandibular canal in relation to an impacted third molar on panoramic radiographs using a segmentation technique. Transfer learning appears to be an effective method for creating such models using a relatively small number of data sets.


Assuntos
Aprendizado Profundo , Canal Mandibular , Dente Serotino , Dente Impactado , Humanos , Canal Mandibular/diagnóstico por imagem , Dente Serotino/diagnóstico por imagem , Radiografia Panorâmica , Dente Impactado/diagnóstico por imagem , Radiografia Dentária Digital
20.
J Contemp Brachytherapy ; 14(1): 87-95, 2022 Feb.
Artigo em Inglês | MEDLINE | ID: mdl-35233240

RESUMO

PURPOSE: The purpose of this study was to evaluate the effect of a lead block for alveolar bone protection in image-guided high-dose-rate interstitial brachytherapy for tongue cancer. MATERIAL AND METHODS: We treated 6 patients and delivered 5,400 cGy in 9 fractions using a lead block. Effects of lead block (median thickness, 4 mm) on dose attenuation by distance were visually examined using TG-43 formalism-based dose distribution curves to determine whether or not the area with the highest dose is located in the alveolar bone, where there is a high-risk of infection. Dose re-calculations were performed using TG-186 formalism with advanced collapsed cone engine (ACE) for inhomogeneity correction set to cortical bone density for the whole mandible and alveolar bone, water density for clinical target volume (CTV), air density for outside body and lead density, and silastic density for lead block and its' silicon replica, respectively. RESULTS: The highest dose was detected outside the alveolar bone in five of the six cases. For dose-volume histogram analysis, median minimum doses delivered per fraction to the 0.1 cm3 of alveolar bone (D0.1cm3 TG-43, ACE-silicon, and ACE-lead) were 344.3 (range, 262.9-427.4) cGy, 336.6 (253.3-425.0) cGy, and 169.7 (114.9-233.3) cGy, respectively. D0.1cm3 ACE-lead was significantly lower than other parameters. No significant difference was observed between CTV-related parameters. CONCLUSIONS: The results suggested that using a lead block for alveolar bone protection with a thickness of about 4 mm, can shift the highest dose area to non-alveolar regions. In addition, it reduced D0.1cm3 of alveolar bone to about half, without affecting tumor dose.

SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA
...