Your browser doesn't support javascript.
loading
Montrer: 20 | 50 | 100
Résultats 1 - 20 de 31
Filtrer
1.
Sci Data ; 11(1): 4, 2024 Jan 02.
Article de Anglais | MEDLINE | ID: mdl-38168517

RÉSUMÉ

Several Diptera species are known to transmit pathogens of medical and veterinary interest. However, identifying these species using conventional methods can be time-consuming, labor-intensive, or expensive. A computer vision-based system that uses Wing interferential patterns (WIPs) to identify these insects could solve this problem. This study introduces a dataset for training and evaluating a recognition system for dipteran insects of medical and veterinary importance using WIPs. The dataset includes pictures of Culicidae, Calliphoridae, Muscidae, Tabanidae, Ceratopogonidae, and Psychodidae. The dataset is complemented by previously published datasets of Glossinidae and some Culicidae members. The new dataset contains 2,399 pictures of 18 genera, with each genus documented by a variable number of species and annotated as a class. The dataset covers species variation, with some genera having up to 300 samples.


Sujet(s)
Ceratopogonidae , Apprentissage profond , Diptera , Muscidae , Animaux , Insectes
2.
J Crohns Colitis ; 18(1): 75-81, 2024 Jan 27.
Article de Anglais | MEDLINE | ID: mdl-37527554

RÉSUMÉ

BACKGROUND AND AIM: Pan-enteric capsule endoscopy [PCE] is a highly sensitive but time-consuming tool for detecting pathology. Artificial intelligence [AI] algorithms might offer a possibility to assist in the review and reduce the analysis time of PCE. This study examines the agreement between PCE assessments aided by AI technology and standard evaluations, in patients suspected of Crohn's disease [CD]. METHOD: PCEs from a prospective, blinded, multicentre study, including patients suspected of CD, were processed by the deep learning solution AXARO® [Augmented Endoscopy, Paris, France]. Based on the image output, two observers classified the patient's PCE as normal or suggestive of CD, ulcerative colitis, or cancer. The primary outcome was per-patient sensitivities and specificities for detecting CD and inflammatory bowel disease [IBD]. Complete reading of PCE served as the reference standard. RESULTS: A total of 131 patients' PCEs were analysed, with a median recording time of 303 min. The AXARO® framework reduced output to a median of 470 images [2.1%] per patient, and the pooled median review time was 3.2 min per patient. For detecting CD, the observers had a sensitivity of 96% and 92% and a specificity of 93% and 90%, respectively. For the detection of IBD, both observers had a sensitivity of 97% and had a specificity of 91% and 90%, respectively. The negative predictive value was 95% for CD and 97% for IBD. CONCLUSIONS: Using the AXARO® framework reduced the initial review time substantially while maintaining high diagnostic accuracy-suggesting its use as a rapid tool to rule out IBD in PCEs of patients suspected of Crohn's disease.


Sujet(s)
Endoscopie par capsule , Maladie de Crohn , Maladies inflammatoires intestinales , Humains , Maladie de Crohn/imagerie diagnostique , Maladie de Crohn/anatomopathologie , Études prospectives , Intelligence artificielle , Maladies inflammatoires intestinales/diagnostic
3.
Sci Rep ; 13(1): 21389, 2023 12 04.
Article de Anglais | MEDLINE | ID: mdl-38049590

RÉSUMÉ

Sandflies (Diptera; Psychodidae) are medical and veterinary vectors that transmit diverse parasitic, viral, and bacterial pathogens. Their identification has always been challenging, particularly at the specific and sub-specific levels, because it relies on examining minute and mostly internal structures. Here, to circumvent such limitations, we have evaluated the accuracy and reliability of Wing Interferential Patterns (WIPs) generated on the surface of sandfly wings in conjunction with deep learning (DL) procedures to assign specimens at various taxonomic levels. Our dataset proves that the method can accurately identify sandflies over other dipteran insects at the family, genus, subgenus, and species level with an accuracy higher than 77.0%, regardless of the taxonomic level challenged. This approach does not require inspection of internal organs to address identification, does not rely on identification keys, and can be implemented under field or near-field conditions, showing promise for sandfly pro-active and passive entomological surveys in an era of scarcity in medical entomologists.


Sujet(s)
Apprentissage profond , Phlebotomus , Psychodidae , Animaux , Psychodidae/parasitologie , Reproductibilité des résultats , Phlebotomus/parasitologie , Entomologie
4.
Biomedicines ; 11(10)2023 Sep 30.
Article de Anglais | MEDLINE | ID: mdl-37893062

RÉSUMÉ

To characterize the growth of brain organoids (BOs), cultures that replicate some early physiological or pathological developments of the human brain are usually manually extracted. Due to their novelty, only small datasets of these images are available, but segmenting the organoid shape automatically with deep learning (DL) tools requires a larger number of images. Light U-Net segmentation architectures, which reduce the training time while increasing the sensitivity under small input datasets, have recently emerged. We further reduce the U-Net architecture and compare the proposed architecture (MU-Net) with U-Net and UNet-Mini on bright-field images of BOs using several data augmentation strategies. In each case, we perform leave-one-out cross-validation on 40 original and 40 synthesized images with an optimized adversarial autoencoder (AAE) or on 40 transformed images. The best results are achieved with U-Net segmentation trained on optimized augmentation. However, our novel method, MU-Net, is more robust: it achieves nearly as accurate segmentation results regardless of the dataset used for training (various AAEs or a transformation augmentation). In this study, we confirm that small datasets of BOs can be segmented with a light U-Net method almost as accurately as with the original method.

5.
Sci Rep ; 13(1): 17628, 2023 10 17.
Article de Anglais | MEDLINE | ID: mdl-37848666

RÉSUMÉ

Hematophagous insects belonging to the Aedes genus are proven vectors of viral and filarial pathogens of medical interest. Aedes albopictus is an increasingly important vector because of its rapid worldwide expansion. In the context of global climate change and the emergence of zoonotic infectious diseases, identification tools with field application are required to strengthen efforts in the entomological survey of arthropods with medical interest. Large scales and proactive entomological surveys of Aedes mosquitoes need skilled technicians and/or costly technical equipment, further puzzled by the vast amount of named species. In this study, we developed an automatic classification system of Aedes species by taking advantage of the species-specific marker displayed by Wing Interferential Patterns. A database holding 494 photomicrographs of 24 Aedes spp. from which those documented with more than ten pictures have undergone a deep learning methodology to train a convolutional neural network and test its accuracy to classify samples at the genus, subgenus, and species taxonomic levels. We recorded an accuracy of 95% at the genus level and > 85% for two (Ochlerotatus and Stegomyia) out of three subgenera tested. Lastly, eight were accurately classified among the 10 Aedes sp. that have undergone a training process with an overall accuracy of > 70%. Altogether, these results demonstrate the potential of this methodology for Aedes species identification and will represent a tool for the future implementation of large-scale entomological surveys.


Sujet(s)
Aedes , Ochlerotatus , Animaux , Vecteurs moustiques , Apprentissage machine , Spécificité d'espèce
6.
Sci Rep ; 13(1): 13895, 2023 08 25.
Article de Anglais | MEDLINE | ID: mdl-37626130

RÉSUMÉ

We present a new and innovative identification method based on deep learning of the wing interferential patterns carried by mosquitoes of the Anopheles genus to classify and assign 20 Anopheles species, including 13 malaria vectors. We provide additional evidence that this approach can identify Anopheles spp. with an accuracy of up to 100% for ten out of 20 species. Although, this accuracy was moderate (> 65%) or weak (50%) for three and seven species. The accuracy of the process to discriminate cryptic or sibling species is also assessed on three species belonging to the Gambiae complex. Strikingly, An. gambiae, An. arabiensis and An. coluzzii, morphologically indistinguishable species belonging to the Gambiae complex, were distinguished with 100%, 100%, and 88% accuracy respectively. Therefore, this tool would help entomological surveys of malaria vectors and vector control implementation. In the future, we anticipate our method can be applied to other arthropod vector-borne diseases.


Sujet(s)
Anopheles , Arthropodes , Apprentissage profond , Animaux , Humains , Vecteurs moustiques , Fratrie
7.
Front Neurosci ; 17: 1220172, 2023.
Article de Anglais | MEDLINE | ID: mdl-37650105

RÉSUMÉ

Introduction: Datasets containing only few images are common in the biomedical field. This poses a global challenge for the development of robust deep-learning analysis tools, which require a large number of images. Generative Adversarial Networks (GANs) are an increasingly used solution to expand small datasets, specifically in the biomedical domain. However, the validation of synthetic images by metrics is still controversial and psychovisual evaluations are time consuming. Methods: We augment a small brain organoid bright-field database of 40 images using several GAN optimizations. We compare these synthetic images to the original dataset using similitude metrcis and we perform an psychovisual evaluation of the 240 images generated. Eight biological experts labeled the full dataset (280 images) as syntetic or natural using a custom-built software. We calculate the error rate per loss optimization as well as the hesitation time. We then compare these results to those provided by the similarity metrics. We test the psychovalidated images in a training step of a segmentation task. Results and discussion: Generated images are considered as natural as the original dataset, with no increase of the hesitation time by experts. Experts are particularly misled by perceptual and Wasserstein loss optimization. These optimizations render the most qualitative and similar images according to metrics to the original dataset. We do not observe a strong correlation but links between some metrics and psychovisual decision according to the kind of generation. Particular Blur metric combinations could maybe replace the psychovisual evaluation. Segmentation task which use the most psychovalidated images are the most accurate.

8.
Article de Anglais | MEDLINE | ID: mdl-37018555

RÉSUMÉ

Anomaly detection is important in many real-life applications. Recently, self-supervised learning has greatly helped deep anomaly detection by recognizing several geometric transformations. However these methods lack finer features, usually highly depend on the anomaly type, and do not perform well on fine-grained problems. To address these issues, we first introduce in this work three novel and efficient discriminative and generative tasks which have complementary strength: (i) a piece-wise jigsaw puzzle task focuses on structure cues; (ii) a tint rotation recognition is used within each piece, taking into account the colorimetry information; (iii) and a partial re-colorization task considers the image texture. In order to make the re-colorization task more object-oriented than background-oriented, we propose to include the contextual color information of the image border via an attention mechanism.We then present a new out-of-distribution detection function and highlight its better stability compared to existing methods. Along with it, we also experiment different score fusion functions. Finally, we evaluate our method on an extensive protocol composed of various anomaly types, from object anomalies, style anomalies with fine-grained classification to local anomalies with face anti-spoofing datasets. Our model significantly outperforms state-of-the-art with up to 36% relative error improvement on object anomalies and 40% on face anti-spoofing problems.

9.
Therap Adv Gastroenterol ; 15: 17562848221132683, 2022.
Article de Anglais | MEDLINE | ID: mdl-36338789

RÉSUMÉ

Background: Artificial intelligence (AI) is rapidly infiltrating multiple areas in medicine, with gastrointestinal endoscopy paving the way in both research and clinical applications. Multiple challenges associated with the incorporation of AI in endoscopy are being addressed in recent consensus documents. Objectives: In the current paper, we aimed to map future challenges and areas of research for the incorporation of AI in capsule endoscopy (CE) practice. Design: Modified three-round Delphi consensus online survey. Methods: The study design was based on a modified three-round Delphi consensus online survey distributed to a group of CE and AI experts. Round one aimed to map out key research statements and challenges for the implementation of AI in CE. All queries addressing the same questions were merged into a single issue. The second round aimed to rank all generated questions during round one and to identify the top-ranked statements with the highest total score. Finally, the third round aimed to redistribute and rescore the top-ranked statements. Results: Twenty-one (16 gastroenterologists and 5 data scientists) experts participated in the survey. In the first round, 48 statements divided into seven themes were generated. After scoring all statements and rescoring the top 12, the question of AI use for identification and grading of small bowel pathologies was scored the highest (mean score 9.15), correlation of AI and human expert reading-second (9.05), and real-life feasibility-third (9.0). Conclusion: In summary, our current study points out a roadmap for future challenges and research areas on our way to fully incorporating AI in CE reading.

10.
Sci Rep ; 12(1): 20086, 2022 11 22.
Article de Anglais | MEDLINE | ID: mdl-36418429

RÉSUMÉ

A simple method for accurately identifying Glossina spp in the field is a challenge to sustain the future elimination of Human African Trypanosomiasis (HAT) as a public health scourge, as well as for the sustainable management of African Animal Trypanosomiasis (AAT). Current methods for Glossina species identification heavily rely on a few well-trained experts. Methodologies that rely on molecular methodologies like DNA barcoding or mass spectrometry protein profiling (MALDI TOFF) haven't been thoroughly investigated for Glossina sp. Nevertheless, because they are destructive, costly, time-consuming, and expensive in infrastructure and materials, they might not be well adapted for the survey of arthropod vectors involved in the transmission of pathogens responsible for Neglected Tropical Diseases, like HAT. This study demonstrates a new type of methodology to classify Glossina species. In conjunction with a deep learning architecture, a database of Wing Interference Patterns (WIPs) representative of the Glossina species involved in the transmission of HAT and AAT was used. This database has 1766 pictures representing 23 Glossina species. This cost-effective methodology, which requires mounting wings on slides and using a commercially available microscope, demonstrates that WIPs are an excellent medium to automatically recognize Glossina species with very high accuracy.


Sujet(s)
Maladie du sommeil , Mouches tsé-tsé , Animaux , Humains , Apprentissage machine , Bases de données factuelles , Maladies négligées , Spectrométrie de masse MALDI
11.
Diagnostics (Basel) ; 12(10)2022 Sep 26.
Article de Anglais | MEDLINE | ID: mdl-36292013

RÉSUMÉ

Capsule endoscopy (CE) is a valid alternative to conventional gastrointestinal (GI) endoscopy tools. In CE, annotation tools are crucial in developing large and annotated medical image databases for training deep neural networks (DNN). We provide an overview of the described and in-use various annotation systems available, focusing on the annotation of adenomatous polyp pathology in the GI tract. Some studies present promising results regarding time efficiency by implementing automated labelling features in annotation systems. Thus, data are inadequate regarding the general overview for users, and may also be more specific on which features provided are necessary for polyp annotation.

12.
J Clin Med ; 11(10)2022 May 17.
Article de Anglais | MEDLINE | ID: mdl-35628947

RÉSUMÉ

Background: Bubbles often mask the mucosa during capsule endoscopy (CE). Clinical scores assessing the cleanliness and the amount of bubbles in the small bowel (SB) are poorly reproducible unlike machine learning (ML) solutions. We aimed to measure the amount of bubbles with ML algorithms in SB CE recordings, and compare two polyethylene glycol (PEG)-based preparations, with and without simethicone, in patients with obscure gastro-intestinal bleeding (OGIB). Patients & Methods: All consecutive outpatients with OGIB from a tertiary care center received a PEG-based preparation, without or with simethicone, in two different periods. The primary outcome was a difference in the proportions (%) of frames with abundant bubbles (>10%) along the full-length video sequences between the two periods. SB CE recordings were analyzed by a validated computed algorithm based on a grey-level of co-occurrence matrix (GLCM), to assess the abundance of bubbles in each frame. Results: In total, 105 third generation SB CE recordings were analyzed (48 without simethicone and 57 with simethicone-added preparations). A significant association was shown between the use of a simethicone-added preparation and a lower abundance of bubbles along the SB (p = 0.04). A significantly lower proportion of "abundant in bubbles" frames was observed in the fourth quartile (30.5% vs. 20.6%, p = 0.02). There was no significant impact of the use of simethicone in terms of diagnostic yield, SB transit time and completion rate. Conclusion: An accurate and reproducible computed algorithm demonstrated significant decrease in the abundance of bubbles along SB CE recordings, with a marked effect in the last quartile, in patients for whom simethicone had been added in PEG-based preparations, compared to those without simethicone.

13.
J Clin Med ; 10(23)2021 Dec 06.
Article de Anglais | MEDLINE | ID: mdl-34884410

RÉSUMÉ

Artificial intelligence (AI) has shown promising results in digestive endoscopy, especially in capsule endoscopy (CE). However, some physicians still have some difficulties and fear the advent of this technology. We aimed to evaluate the perceptions and current sentiments toward the use of AI in CE. An online survey questionnaire was sent to an audience of gastroenterologists. In addition, several European national leaders of the International CApsule endoscopy REsearch (I CARE) Group were asked to disseminate an online survey among their national communities of CE readers (CER). The survey included 32 questions regarding general information, perceptions of AI, and its use in daily life, medicine, endoscopy, and CE. Among 380 European gastroenterologists who answered this survey, 333 (88%) were CERs. The mean average time length of experience in CE reading was 9.9 years (0.5-22). A majority of CERs agreed that AI would positively impact CE, shorten CE reading time, and help standardize reporting in CE and characterize lesions seen in CE. Nevertheless, in the foreseeable future, a majority of CERs disagreed with the complete replacement all CE reading by AI. Most CERs believed in the high potential of AI for becoming a valuable tool for automated diagnosis and for shortening the reading time. Currently, the perception is that AI will not replace CE reading.

14.
Dig Liver Dis ; 53(12): 1627-1631, 2021 Dec.
Article de Anglais | MEDLINE | ID: mdl-34563469

RÉSUMÉ

BACKGROUND AND AIMS: Current artificial intelligence (AI)-based solutions for capsule endoscopy (CE) interpretation are proprietary. We aimed to evaluate an AI solution trained on a specific CE system (Pillcam®, Medtronic) for the detection of angiectasias on images captured by a different proprietary system (MiroCam®, Intromedic). MATERIAL AND METHODS: An advanced AI solution (Axaro®, Augmented Endoscopy), previously trained on Pillcam® small bowell images, was evaluated on independent datasets with more than 1200 Pillcam® and MiroCam® still frames (equally distributed, with or without angiectasias). Images were reviewed by experts before and after AI interpretation. RESULTS: Sensitivity for the diagnosis of angiectasia was 97.4% with Pillcam® images and 96.1% with Mirocam® images, with specificity of 98.8% and 97.8%, respectively. Performances regarding the delineation of regions of interest and the characterization of angiectasias were similar in both groups (all above 95%). Processing time was significantly shorter with Mirocam® (20.7 ms) than with Pillcam® images (24.6 ms, p<0.0001), possibly related to technical differences between systems. CONCLUSION: This proof-of-concept study on still images paves the way for the development of resource-sparing, "universal" CE databases and AI solutions for CE interpretation.


Sujet(s)
Angiodysplasie/diagnostic , Endoscopie par capsule/méthodes , Apprentissage profond , Intestin grêle/anatomopathologie , Humains , Intestin grêle/imagerie diagnostique , Étude de validation de principe
15.
J Imaging ; 7(2)2021 Feb 03.
Article de Anglais | MEDLINE | ID: mdl-34460624

RÉSUMÉ

Bio-inspired Event-Based (EB) cameras are a promising new technology that outperforms standard frame-based cameras in extreme lighted and fast moving scenes. Already, a number of EB corner detection techniques have been developed; however, the performance of these EB corner detectors has only been evaluated based on a few author-selected criteria rather than on a unified common basis, as proposed here. Moreover, their experimental conditions are mainly limited to less interesting operational regions of the EB camera (on which frame-based cameras can also operate), and some of the criteria, by definition, could not distinguish if the detector had any systematic bias. In this paper, we evaluate five of the seven existing EB corner detectors on a public dataset including extreme illumination conditions that have not been investigated before. Moreover, this evaluation is the first of its kind in terms of analysing not only such a high number of detectors, but also applying a unified procedure for all. Contrary to previous assessments, we employed both the intensity and trajectory information within the public dataset rather than only one of them. We show that a rigorous comparison among EB detectors can be performed without tedious manual labelling and even with challenging acquisition conditions. This study thus proposes the first standard unified EB corner evaluation procedure, which will enable better understanding of the underlying mechanisms of EB cameras and can therefore lead to more efficient EB corner detection techniques.

16.
Front Neurosci ; 15: 629067, 2021.
Article de Anglais | MEDLINE | ID: mdl-34276279

RÉSUMÉ

Purpose: Since their first generation in 2013, the use of cerebral organoids has spread exponentially. Today, the amount of generated data is becoming challenging to analyze manually. This review aims to overview the current image acquisition methods and to subsequently identify the needs in image analysis tools for cerebral organoids. Methods: To address this question, we went through all recent articles published on the subject and annotated the protocols, acquisition methods, and algorithms used. Results: Over the investigated period of time, confocal microscopy and bright-field microscopy were the most used acquisition techniques. Cell counting, the most common task, is performed in 20% of the articles and area; around 12% of articles calculate morphological parameters. Image analysis on cerebral organoids is performed in majority using ImageJ software (around 52%) and Matlab language (4%). Treatments remain mostly semi-automatic. We highlight the limitations encountered in image analysis in the cerebral organoid field and suggest possible solutions and implementations to develop. Conclusions: In addition to providing an overview of cerebral organoids cultures and imaging, this work highlights the need to improve the existing image analysis methods for such images and the need for specific analysis tools. These solutions could specifically help to monitor the growth of future standardized cerebral organoids.

17.
J Gastroenterol Hepatol ; 36(1): 12-19, 2021 Jan.
Article de Anglais | MEDLINE | ID: mdl-33448511

RÉSUMÉ

Neural network-based solutions are under development to alleviate physicians from the tedious task of small-bowel capsule endoscopy reviewing. Computer-assisted detection is a critical step, aiming to reduce reading times while maintaining accuracy. Weakly supervised solutions have shown promising results; however, video-level evaluations are scarce, and no prospective studies have been conducted yet. Automated characterization (in terms of diagnosis and pertinence) by supervised machine learning solutions is the next step. It relies on large, thoroughly labeled databases, for which preliminary "ground truth" definitions by experts are of tremendous importance. Other developments are under ways, to assist physicians in localizing anatomical landmarks and findings in the small bowel, in measuring lesions, and in rating bowel cleanliness. It is still questioned whether artificial intelligence will enter the market with proprietary, built-in or plug-in software, or with a universal cloud-based service, and how it will be accepted by physicians and patients.


Sujet(s)
Intelligence artificielle/tendances , Endoscopie par capsule/méthodes , Endoscopie par capsule/tendances , Maladies intestinales/diagnostic , Maladies intestinales/anatomopathologie , Intestin grêle/anatomopathologie , Prévision , Humains
18.
Endoscopy ; 53(9): 932-936, 2021 09.
Article de Anglais | MEDLINE | ID: mdl-33137834

RÉSUMÉ

BACKGROUND: Cleanliness scores in small-bowel capsule endoscopy (SBCE) have poor reproducibility. The aim of this study was to evaluate a neural network-based algorithm for automated assessment of small-bowel cleanliness during capsule endoscopy. METHODS: 600 normal third-generation SBCE still frames were categorized as "adequate" or "inadequate" in terms of cleanliness by three expert readers, according to a 10-point scale, and served as a training database. Then, 156 third-generation SBCE recordings were categorized in a consensual manner as "adequate" or "inadequate" in terms of cleanliness; this testing database was split into two independent 78-video subsets for the tuning and evaluation of the algorithm, respectively. RESULTS: Using a threshold of 79 % "adequate" still frames per video to achieve the best performance, the algorithm yielded a sensitivity of 90.3 %, specificity of 83.3 %, and accuracy of 89.7 %. The reproducibility was perfect. The mean calculation time per video was 3 (standard deviation 1) minutes. CONCLUSION: This neural network-based algorithm allowing automatic assessment of small-bowel cleanliness during capsule endoscopy was highly sensitive and paves the way for automated, standardized SBCE reports.


Sujet(s)
Endoscopie par capsule , Algorithmes , Humains , Intestin grêle/imagerie diagnostique , , Reproductibilité des résultats
19.
Endoscopy ; 53(9): 893-901, 2021 09.
Article de Anglais | MEDLINE | ID: mdl-33167043

RÉSUMÉ

BACKGROUND : Artificial intelligence (AI) research in colonoscopy is progressing rapidly but widespread clinical implementation is not yet a reality. We aimed to identify the top implementation research priorities. METHODS : An established modified Delphi approach for research priority setting was used. Fifteen international experts, including endoscopists and translational computer scientists/engineers, from nine countries participated in an online survey over 9 months. Questions related to AI implementation in colonoscopy were generated as a long-list in the first round, and then scored in two subsequent rounds to identify the top 10 research questions. RESULTS : The top 10 ranked questions were categorized into five themes. Theme 1: clinical trial design/end points (4 questions), related to optimum trial designs for polyp detection and characterization, determining the optimal end points for evaluation of AI, and demonstrating impact on interval cancer rates. Theme 2: technological developments (3 questions), including improving detection of more challenging and advanced lesions, reduction of false-positive rates, and minimizing latency. Theme 3: clinical adoption/integration (1 question), concerning the effective combination of detection and characterization into one workflow. Theme 4: data access/annotation (1 question), concerning more efficient or automated data annotation methods to reduce the burden on human experts. Theme 5: regulatory approval (1 question), related to making regulatory approval processes more efficient. CONCLUSIONS : This is the first reported international research priority setting exercise for AI in colonoscopy. The study findings should be used as a framework to guide future research with key stakeholders to accelerate the clinical implementation of AI in endoscopy.


Sujet(s)
Intelligence artificielle , Coloscopie , Méthode Delphi , Humains
20.
Endosc Int Open ; 8(3): E415-E420, 2020 Mar.
Article de Anglais | MEDLINE | ID: mdl-32118115

RÉSUMÉ

Background and study aims Capsule endoscopy (CE) is the preferred method for small bowel (SB) exploration. With a mean number of 50,000 SB frames per video, SBCE reading is time-consuming and tedious (30 to 60 minutes per video). We describe a large, multicenter database named CAD-CAP (Computer-Assisted Diagnosis for CAPsule Endoscopy, CAD-CAP). This database aims to serve the development of CAD tools for CE reading. Materials and methods Twelve French endoscopy centers were involved. All available third-generation SB-CE videos (Pillcam, Medtronic) were retrospectively selected from these centers and deidentified. Any pathological frame was extracted and included in the database. Manual segmentation of findings within these frames was performed by two pre-med students trained and supervised by an expert reader. All frames were then classified by type and clinical relevance by a panel of three expert readers. An automated extraction process was also developed to create a dataset of normal, proofread, control images from normal, complete, SB-CE videos. Results Four-thousand-one-hundred-and-seventy-four SB-CE were included. Of them, 1,480 videos (35 %) containing at least one pathological finding were selected. Findings from 5,184 frames (with their short video sequences) were extracted and delimited: 718 frames with fresh blood, 3,097 frames with vascular lesions, and 1,369 frames with inflammatory and ulcerative lesions. Twenty-thousand normal frames were extracted from 206 SB-CE normal videos. CAD-CAP has already been used for development of automated tools for angiectasia detection and also for two international challenges on medical computerized analysis.

SÉLECTION CITATIONS
DÉTAIL DE RECHERCHE