RESUMEN
Reducing the dose of radiation in computed tomography (CT) is vital to decreasing secondary cancer risk. However, the use of low-dose CT (LDCT) images is accompanied by increased noise that can negatively impact diagnoses. Although numerous deep learning algorithms have been developed for LDCT denoising, several challenges persist, including the visual incongruence experienced by radiologists, unsatisfactory performances across various metrics, and insufficient exploration of the networks' robustness in other CT domains. To address such issues, this study proposes three novel accretions. First, we propose a generative adversarial network (GAN) with a robust discriminator through multi-task learning that simultaneously performs three vision tasks: restoration, image-level, and pixel-level decisions. The more multi-tasks that are performed, the better the denoising performance of the generator, which means multi-task learning enables the discriminator to provide more meaningful feedback to the generator. Second, two regulatory mechanisms, restoration consistency (RC) and non-difference suppression (NDS), are introduced to improve the discriminator's representation capabilities. These mechanisms eliminate irrelevant regions and compare the discriminator's results from the input and restoration, thus facilitating effective GAN training. Lastly, we incorporate residual fast Fourier transforms with convolution (Res-FFT-Conv) blocks into the generator to utilize both frequency and spatial representations. This approach provides mixed receptive fields by using spatial (or local), spectral (or global), and residual connections. Our model was evaluated using various pixel- and feature-space metrics in two denoising tasks. Additionally, we conducted visual scoring with radiologists. The results indicate superior performance in both quantitative and qualitative measures compared to state-of-the-art denoising techniques.
RESUMEN
BACKGROUND AND AIM: Reliable bowel preparation assessment is important in colonoscopy. However, current scoring systems are limited by laborious and time-consuming tasks and interobserver variability. We aimed to develop an artificial intelligence (AI) model to assess bowel cleanliness and evaluate its clinical applicability. METHODS: A still image-driven AI model to assess the Boston Bowel Preparation Scale (BBPS) was developed and validated using 2361 colonoscopy images. For evaluating real-world applicability, the model was validated using 113 10-s colonoscopy video clips and 30 full colonoscopy videos to identify "adequate (BBPS 2-3)" or "inadequate (BBPS 0-1)" preparation. The model was tested with an external dataset of 29 colonoscopy videos. The clinical applicability of the model was evaluated using 225 consecutive colonoscopies. Inter-rater variability was analyzed between the AI model and endoscopists. RESULTS: The AI model achieved an accuracy of 94.0% and an area under the receiver operating characteristic curve of 0.939 with the still images. Model testing with an external dataset showed an accuracy of 95.3%, an area under the receiver operating characteristic curve of 0.976, and a sensitivity of 100% for the detection of inadequate preparations. The clinical applicability study showed an overall agreement rate of 85.3% between endoscopists and the AI model, with Fleiss' kappa of 0.686. The agreement rate was lower for the right colon compared with the transverse and left colon, with Fleiss' kappa of 0.563, 0.575, and 0.789, respectively. CONCLUSIONS: The AI model demonstrated accurate bowel preparation assessment and substantial agreement with endoscopists. Further refinement of the AI model is warranted for effective monitoring of qualified colonoscopy in large-scale screening programs.
Asunto(s)
Inteligencia Artificial , Colonoscopía , Humanos , Colonoscopía/métodos , Femenino , Persona de Mediana Edad , Masculino , Catárticos/administración & dosificación , Anciano , Variaciones Dependientes del Observador , Adulto , Curva ROCRESUMEN
BACKGROUND AND OBJECTIVE: Despite recent development of AI, prediction of the surgical movement in the maxilla and mandible by OGS might be more difficult than that of tooth movement by orthodontic treatment. To evaluate the prediction accuracy of the surgical movement using pairs of pre-(T0) and post-surgical (T1) lateral cephalograms (lat-ceph) of orthognathic surgery (OGS) patients and dual embedding module-graph convolution neural network (DEM-GCNN) model. METHODS: 599 pairs from 3 institutions were used as training, internal validation, and internal test sets and 201 pairs from other 6 institutions were used as external test set. DEM-GCNN model (IEM, learning the lat-ceph images; LTEM, learning the landmarks) was developed to predict the amount and direction of surgical movement of ANS and PNS in the maxilla and B-point and Md1crown in the mandible. The distance between T1 landmark coordinates actually moved by OGS (ground truth) and predicted by DEM-GCNN model and pre-existed CNN-based Model-C (learning the lat-ceph images) was compared. RESULTS: In both internal and external tests, DEM-GCNN did not exhibit significant difference from ground truth in all landmarks (ANS, PNS, B-point, Md1crown, all P > 0.05). When the accumulated successful detection rate for each landmark was compared, DEM-GCNN showed higher values than Model-C in both the internal and external tests. In violin plots exhibiting the error distribution of the prediction results, both internal and external tests showed that DEM-GCNN had significant performance improvement in PNS, ANS, B-point, Md1crown than Model-C. DEM-GCNN showed significantly lower prediction error values than Model-C (one-jaw surgery, B-point, Md1crown, all P < 0.005; two-jaw surgery, PNS, ANS, all P < 0.05; B point, Md1crown, all P < 0.005). CONCLUSION: We developed a robust OGS planning model with maximized generalizability despite diverse qualities of lat-cephs from 9 institutions.