Your browser doesn't support javascript.
loading
: 20 | 50 | 100
1 - 20 de 34.415
1.
Sci Rep ; 14(1): 13244, 2024 06 09.
Article En | MEDLINE | ID: mdl-38853158

Aiming at the problem of image classification with insignificant morphological structural features, strong target correlation, and low signal-to-noise ratio, combined with prior feature knowledge embedding, a deep learning method based on ResNet and Radial Basis Probabilistic Neural Network (RBPNN) is proposed model. Taking ResNet50 as a visual modeling network, it uses feature pyramid and self-attention mechanism to extract appearance and semantic features of images at multiple scales, and associate and enhance local and global features. Taking into account the diversity of category features, channel cosine similarity attention and dynamic C-means clustering algorithms are used to select representative sample features in different category of sample subsets to implicitly express prior category feature knowledge, and use them as the kernel centers of radial basis probability neurons (RBPN) to realize the embedding of diverse prior feature knowledge. In the RBPNN pattern aggregation layer, the outputs of RBPN are selectively summed according to the category of the kernel center, that is, the subcategory features are combined into category features, and finally the image classification is implemented based on Softmax. The functional module of the proposed method is designed specifically for image characteristics, which can highlight the significance of local and structural features of the image, form a non-convex decision-making area, and reduce the requirements for the completeness of the sample set. Applying the proposed method to medical image classification, experiments were conducted based on the brain tumor MRI image classification public dataset and the actual cardiac ultrasound image dataset, and the accuracy rate reached 85.82% and 83.92% respectively. Compared with the three mainstream image classification models, the performance indicators of this method have been significantly improved.


Deep Learning , Neural Networks, Computer , Humans , Algorithms , Image Processing, Computer-Assisted/methods , Brain Neoplasms/diagnostic imaging , Brain Neoplasms/classification , Brain Neoplasms/pathology , Image Interpretation, Computer-Assisted/methods , Magnetic Resonance Imaging/methods
2.
Comput Methods Programs Biomed ; 253: 108238, 2024 Aug.
Article En | MEDLINE | ID: mdl-38823117

BACKGROUND AND OBJECTIVE: Evaluating the interpretability of Deep Learning models is crucial for building trust and gaining insights into their decision-making processes. In this work, we employ class activation map based attribution methods in a setting where only High-Resolution Class Activation Mapping (HiResCAM) is known to produce faithful explanations. The objective is to evaluate the quality of the attribution maps using quantitative metrics and investigate whether faithfulness aligns with the metrics results. METHODS: We fine-tune pre-trained deep learning architectures over four medical image datasets in order to calculate attribution maps. The maps are evaluated on a threefold metrics basis utilizing well-established evaluation scores. RESULTS: Our experimental findings suggest that the Area Over Perturbation Curve (AOPC) and Max-Sensitivity scores favor the HiResCAM maps. On the other hand, the Heatmap Assisted Accuracy Score (HAAS) does not provide insights to our comparison as it evaluates almost all maps as inaccurate. To this purpose we further compare our calculated values against values obtained over a diverse group of models which are trained on non-medical benchmark datasets, to eventually achieve more responsive results. CONCLUSION: This study develops a series of experiments to discuss the connection between faithfulness and quantitative metrics over medical attribution maps. HiResCAM preserves the gradient effect on a pixel level ultimately producing high-resolution, informative and resilient mappings. In turn, this is depicted in the results of AOPC and Max-Sensitivity metrics, successfully identifying the faithful algorithm. In regards to HAAS, our experiments yield that it is sensitive over complex medical patterns, commonly characterized by strong color dependency and multiple attention areas.


Deep Learning , Humans , Algorithms , Diagnostic Imaging , Image Processing, Computer-Assisted/methods , Image Interpretation, Computer-Assisted/methods , Neural Networks, Computer
3.
Math Biosci Eng ; 21(4): 5092-5117, 2024 Mar 04.
Article En | MEDLINE | ID: mdl-38872528

Glaucoma is a chronic neurodegenerative disease that can result in irreversible vision loss if not treated in its early stages. The cup-to-disc ratio is a key criterion for glaucoma screening and diagnosis, and it is determined by dividing the area of the optic cup (OC) by that of the optic disc (OD) in fundus images. Consequently, the automatic and accurate segmentation of the OC and OD is a pivotal step in glaucoma detection. In recent years, numerous methods have resulted in great success on this task. However, most existing methods either have unsatisfactory segmentation accuracy or high time costs. In this paper, we propose a lightweight deep-learning architecture for the simultaneous segmentation of the OC and OD, where we have adopted fuzzy learning and a multi-layer perceptron to simplify the learning complexity and improve segmentation accuracy. Experimental results demonstrate the superiority of our proposed method as compared to most state-of-the-art approaches in terms of both training time and segmentation accuracy.


Algorithms , Deep Learning , Fuzzy Logic , Glaucoma , Optic Disk , Humans , Optic Disk/diagnostic imaging , Glaucoma/diagnostic imaging , Neural Networks, Computer , Image Processing, Computer-Assisted/methods , Reproducibility of Results , Image Interpretation, Computer-Assisted/methods , Fundus Oculi
4.
BMC Med Imaging ; 24(1): 133, 2024 Jun 05.
Article En | MEDLINE | ID: mdl-38840240

BACKGROUND: Breast cancer is the most common cancer among women, and ultrasound is a usual tool for early screening. Nowadays, deep learning technique is applied as an auxiliary tool to provide the predictive results for doctors to decide whether to make further examinations or treatments. This study aimed to develop a hybrid learning approach for breast ultrasound classification by extracting more potential features from local and multi-center ultrasound data. METHODS: We proposed a hybrid learning approach to classify the breast tumors into benign and malignant. Three multi-center datasets (BUSI, BUS, OASBUD) were used to pretrain a model by federated learning, then every dataset was fine-tuned at local. The proposed model consisted of a convolutional neural network (CNN) and a graph neural network (GNN), aiming to extract features from images at a spatial level and from graphs at a geometric level. The input images are small-sized and free from pixel-level labels, and the input graphs are generated automatically in an unsupervised manner, which saves the costs of labor and memory space. RESULTS: The classification AUCROC of our proposed method is 0.911, 0.871 and 0.767 for BUSI, BUS and OASBUD. The balanced accuracy is 87.6%, 85.2% and 61.4% respectively. The results show that our method outperforms conventional methods. CONCLUSIONS: Our hybrid approach can learn the inter-feature among multi-center data and the intra-feature of local data. It shows potential in aiding doctors for breast tumor classification in ultrasound at an early stage.


Breast Neoplasms , Deep Learning , Neural Networks, Computer , Ultrasonography, Mammary , Humans , Breast Neoplasms/diagnostic imaging , Female , Ultrasonography, Mammary/methods , Image Interpretation, Computer-Assisted/methods , Adult
5.
Clin Oral Investig ; 28(7): 364, 2024 Jun 08.
Article En | MEDLINE | ID: mdl-38849649

OBJECTIVES: Diagnosing oral potentially malignant disorders (OPMD) is critical to prevent oral cancer. This study aims to automatically detect and classify the most common pre-malignant oral lesions, such as leukoplakia and oral lichen planus (OLP), and distinguish them from oral squamous cell carcinomas (OSCC) and healthy oral mucosa on clinical photographs using vision transformers. METHODS: 4,161 photographs of healthy mucosa, leukoplakia, OLP, and OSCC were included. Findings were annotated pixel-wise and reviewed by three clinicians. The photographs were divided into 3,337 for training and validation and 824 for testing. The training and validation images were further divided into five folds with stratification. A Mask R-CNN with a Swin Transformer was trained five times with cross-validation, and the held-out test split was used to evaluate the model performance. The precision, F1-score, sensitivity, specificity, and accuracy were calculated. The area under the receiver operating characteristics curve (AUC) and the confusion matrix of the most effective model were presented. RESULTS: The detection of OSCC with the employed model yielded an F1 of 0.852 and AUC of 0.974. The detection of OLP had an F1 of 0.825 and AUC of 0.948. For leukoplakia the F1 was 0.796 and the AUC was 0.938. CONCLUSIONS: OSCC were effectively detected with the employed model, whereas the detection of OLP and leukoplakia was moderately effective. CLINICAL RELEVANCE: Oral cancer is often detected in advanced stages. The demonstrated technology may support the detection and observation of OPMD to lower the disease burden and identify malignant oral cavity lesions earlier.


Leukoplakia, Oral , Lichen Planus, Oral , Mouth Neoplasms , Precancerous Conditions , Humans , Mouth Neoplasms/diagnosis , Precancerous Conditions/diagnosis , Lichen Planus, Oral/diagnosis , Leukoplakia, Oral/diagnosis , Sensitivity and Specificity , Photography , Diagnosis, Differential , Carcinoma, Squamous Cell/diagnosis , Male , Female , Photography, Dental , Image Interpretation, Computer-Assisted/methods
6.
Math Biosci Eng ; 21(4): 5250-5282, 2024 Mar 06.
Article En | MEDLINE | ID: mdl-38872535

The increasing global incidence of glioma tumors has raised significant healthcare concerns due to their high mortality rates. Traditionally, tumor diagnosis relies on visual analysis of medical imaging and invasive biopsies for precise grading. As an alternative, computer-assisted methods, particularly deep convolutional neural networks (DCNNs), have gained traction. This research paper explores the recent advancements in DCNNs for glioma grading using brain magnetic resonance images (MRIs) from 2015 to 2023. The study evaluated various DCNN architectures and their performance, revealing remarkable results with models such as hybrid and ensemble based DCNNs achieving accuracy levels of up to 98.91%. However, challenges persisted in the form of limited datasets, lack of external validation, and variations in grading formulations across diverse literature sources. Addressing these challenges through expanding datasets, conducting external validation, and standardizing grading formulations can enhance the performance and reliability of DCNNs in glioma grading, thereby advancing brain tumor classification and extending its applications to other neurological disorders.


Brain Neoplasms , Deep Learning , Glioma , Magnetic Resonance Imaging , Neoplasm Grading , Neural Networks, Computer , Humans , Glioma/diagnostic imaging , Glioma/pathology , Glioma/classification , Magnetic Resonance Imaging/methods , Brain Neoplasms/diagnostic imaging , Brain Neoplasms/pathology , Reproducibility of Results , Algorithms , Brain/diagnostic imaging , Brain/pathology , Image Processing, Computer-Assisted/methods , Image Interpretation, Computer-Assisted/methods
7.
Diagn Pathol ; 19(1): 75, 2024 Jun 08.
Article En | MEDLINE | ID: mdl-38851736

BACKGROUND & OBJECTIVES: Tumor grade determines prognosis in urothelial carcinoma. The classification of low and high grade is based on nuclear morphological features that include nuclear size, hyperchromasia and pleomorphism. These features are subjectively assessed by the pathologists and are not numerically measured, which leads to high rates of interobserver variability. The purpose of this study is to assess the value of a computer-based image analysis tool for identifying predictors of tumor grade in bladder cancer. METHODS: Four hundred images of urothelial tumors were graded by five pathologists and two expert genitourinary pathologists using a scale of 1 (lowest grade) to 5 (highest grade). A computer algorithm was used to automatically segment the nuclei and to provide morphometric parameters for each nucleus, which were used to establish the grading algorithm. Grading algorithm was compared to pathologists' agreement. RESULTS: Comparison of the grading scores of the five pathologists with the expert genitourinary pathologists score showed agreement rates between 88.5% and 97.5%.The agreement rate between the two expert genitourinary pathologists was 99.5%. The quantified algorithm based conventional parameters that determine the grade (nuclear size, pleomorphism and hyperchromasia) showed > 85% agreement with the expert genitourinary pathologists. Surprisingly, the parameter that was most associated with tumor grade was the 10th percentile of the nuclear area, and high grade was associated with lower 10th percentile nuclei, caused by the presence of more inflammatory cells in the high-grade tumors. CONCLUSION: Quantitative nuclear features could be applied to determine urothelial carcinoma grade and explore new biologically explainable parameters with better correlation to grade than those currently used.


Algorithms , Cell Nucleus , Neoplasm Grading , Urinary Bladder Neoplasms , Humans , Urinary Bladder Neoplasms/pathology , Neoplasm Grading/methods , Cell Nucleus/pathology , Observer Variation , Image Interpretation, Computer-Assisted/methods , Image Processing, Computer-Assisted/methods , Carcinoma, Transitional Cell/pathology
8.
Comput Methods Programs Biomed ; 253: 108237, 2024 Aug.
Article En | MEDLINE | ID: mdl-38820715

BACKGROUND AND OBJECTIVES: Graph neural network (GNN) has been extensively used in histopathology whole slide image (WSI) analysis due to the efficiency and flexibility in modelling relationships among entities. However, most existing GNN-based WSI analysis methods only consider the pairwise correlation of patches from one single perspective (e.g. spatial affinity or embedding similarity) yet ignore the intrinsic non-pairwise relationships present in gigapixel WSI, which are likely to contribute to feature learning and downstream tasks. The objective of this study is therefore to explore the non-pairwise relationships in histopathology WSI and exploit them to guide the learning of slide-level representations for better classification performance. METHODS: In this paper, we propose a novel Masked HyperGraph Learning (MaskHGL) framework for weakly supervised histopathology WSI classification. Compared with most GNN-based WSI classification methods, MaskHGL exploits the non-pairwise correlations between patches with hypergraph and global message passing conducted by hypergraph convolution. Concretely, multi-perspective hypergraphs are first built for each WSI, then hypergraph attention is introduced into the jointed hypergraph to propagate the non-pairwise relationships and thus yield more discriminative node representation. More importantly, a masked hypergraph reconstruction module is devised to guide the hypergraph learning which can generate more powerful robustness and generalization than the method only using hypergraph modelling. Additionally, a self-attention-based node aggregator is also applied to explore the global correlation of patches in WSI and produce the slide-level representation for classification. RESULTS: The proposed method is evaluated on two public TCGA benchmark datasets and one in-house dataset. On the public TCGA-LUNG (1494 WSIs) and TCGA-EGFR (696 WSIs) test set, the area under receiver operating characteristic (ROC) curve (AUC) were 0.9752±0.0024 and 0.7421±0.0380, respectively. On the USTC-EGFR (754 WSIs) dataset, MaskHGL achieved significantly better performance with an AUC of 0.8745±0.0100, which surpassed the second-best state-of-the-art method SlideGraph+ 2.64%. CONCLUSIONS: MaskHGL shows a great improvement, brought by considering the intrinsic non-pairwise relationships within WSI, in multiple downstream WSI classification tasks. In particular, the designed masked hypergraph reconstruction module promisingly alleviates the data scarcity and greatly enhances the robustness and classification ability of our MaskHGL. Notably, it has shown great potential in cancer subtyping and fine-grained lung cancer gene mutation prediction from hematoxylin and eosin (H&E) stained WSIs.


Neural Networks, Computer , Humans , Algorithms , Supervised Machine Learning , Image Processing, Computer-Assisted/methods , Lung Neoplasms/pathology , Image Interpretation, Computer-Assisted/methods
9.
Technol Health Care ; 32(S1): 277-286, 2024.
Article En | MEDLINE | ID: mdl-38759056

BACKGROUND: Early diagnosis of knee osteoarthritis is an important area of research in the field of clinical medicine. Due to the complexity in the MRI imaging sequences and the diverse structure of cartilage, there are many challenges in the segmentation of knee bone and cartilage. Relevant studies have conducted semantic fusion processing through splicing or summing forms, which results in reduced resolution and the accumulation of redundant information. OBJECTIVE: This study was envisaged to construct an MRI image segmentation model to improve the diagnostic efficiency and accuracy of different grade knee osteoarthritis by adopting the Dual Attention and Multi-scale Feature Fusion Segmentation network (DA-MFFSnet). METHODS: The feature information of different scales was fused through the Multi-scale Attention Downsample module to extract more accurate feature information, and the Global Attention Upsample module weighted lower-level feature information to reduce the loss of key information. RESULTS: The collected MRI knee images were screened and labeled, and the study results showed that the segmentation effect of DA-MFFSNet model was closer to that of the manually labeled images. The mean intersection over union, the dice similarity coefficient and the volumetric overlap error was 92.74%, 91.08% and 7.44%, respectively, and the accuracy of the differential diagnosis of knee osteoarthritis was 84.42%. CONCLUSIONS: The model exhibited better stability and classification effect. Our results indicated that the Dual Attention and Multi-scale Feature Fusion Segmentation model can improve the segmentation effect of MRI knee images in mild and medium knee osteoarthritis, thereby offering an important clinical value and improving the accuracy of the clinical diagnosis.


Magnetic Resonance Imaging , Osteoarthritis, Knee , Humans , Magnetic Resonance Imaging/methods , Osteoarthritis, Knee/diagnostic imaging , Knee Joint/diagnostic imaging , Image Processing, Computer-Assisted/methods , Algorithms , Image Interpretation, Computer-Assisted/methods
10.
Technol Health Care ; 32(S1): 465-475, 2024.
Article En | MEDLINE | ID: mdl-38759069

BACKGROUND: Oral cancer is a malignant tumor that usually occurs within the tissues of the mouth. This type of cancer mainly includes tumors in the lining of the mouth, tongue, lips, buccal mucosa and gums. Oral cancer is on the rise globally, especially in some specific risk groups. The early stage of oral cancer is usually asymptomatic, while the late stage may present with ulcers, lumps, bleeding, etc. OBJECTIVE: The objective of this paper is to propose an effective and accurate method for the identification and classification of oral cancer. METHODS: We applied two deep learning methods, CNN and Transformers. First, we propose a new CANet classification model for oral cancer, which uses attention mechanisms combined with neglected location information to explore the complex combination of attention mechanisms and deep networks, and fully tap the potential of attention mechanisms. Secondly, we design a classification model based on Swim transform. The image is segmented into a series of two-dimensional image blocks, which are then processed by multiple layers of conversion blocks. RESULTS: The proposed classification model was trained and predicted on Kaggle Oral Cancer Images Dataset, and satisfactory results were obtained. The average accuracy, sensitivity, specificity and F1-Socre of Swin transformer architecture are 94.95%, 95.37%, 95.52% and 94.66%, respectively. The average accuracy, sensitivity, specificity and F1-Score of CANet model were 97.00%, 97.82%, 97.82% and 96.61%, respectively. CONCLUSIONS: We studied different deep learning algorithms for oral cancer classification, including convolutional neural networks, converters, etc. Our Attention module in CANet leverages the benefits of channel attention to model the relationships between channels while encoding precise location information that captures the long-term dependencies of the network. The model achieves a high classification effect with an accuracy of 97.00%, which can be used in the automatic recognition and classification of oral cancer.


Deep Learning , Mouth Neoplasms , Mouth Neoplasms/diagnostic imaging , Mouth Neoplasms/classification , Mouth Neoplasms/pathology , Mouth Neoplasms/diagnosis , Humans , Neural Networks, Computer , Sensitivity and Specificity , Image Interpretation, Computer-Assisted/methods , Image Processing, Computer-Assisted/methods
11.
Comput Biol Med ; 176: 108564, 2024 Jun.
Article En | MEDLINE | ID: mdl-38744010

Alzheimer's disease (AD) is a progressive neurodegenerative condition, and early intervention can help slow its progression. However, integrating multi-dimensional information and deep convolutional networks increases the model parameters, affecting diagnosis accuracy and efficiency and hindering clinical diagnostic model deployment. Multi-modal neuroimaging can offer more precise diagnostic results, while multi-task modeling of classification and regression tasks can enhance the performance and stability of AD diagnosis. This study proposes a Hierarchical Attention-based Multi-task Multi-modal Fusion model (HAMMF) that leverages multi-modal neuroimaging data to concurrently learn AD classification tasks, cognitive score regression, and age regression tasks using attention-based techniques. Firstly, we preprocess MRI and PET image data to obtain two modal data, each containing distinct information. Next, we incorporate a novel Contextual Hierarchical Attention Module (CHAM) to aggregate multi-modal features. This module employs channel and spatial attention to extract fine-grained pathological features from unimodal image data across various dimensions. Using these attention mechanisms, the Transformer can effectively capture correlated features of multi-modal inputs. Lastly, we adopt multi-task learning in our model to investigate the influence of different variables on diagnosis, with a primary classification task and a secondary regression task for optimal multi-task prediction performance. Our experiments utilized MRI and PET images from 720 subjects in the Alzheimer's Disease Neuroimaging Initiative (ADNI) dataset. The results show that our proposed model achieves an overall accuracy of 93.15% for AD/NC recognition, and the visualization results demonstrate its strong pathological feature recognition performance.


Alzheimer Disease , Magnetic Resonance Imaging , Alzheimer Disease/diagnostic imaging , Humans , Magnetic Resonance Imaging/methods , Neuroimaging/methods , Diagnosis, Computer-Assisted/methods , Male , Positron-Emission Tomography/methods , Female , Aged , Brain/diagnostic imaging , Image Interpretation, Computer-Assisted/methods
12.
Comput Biol Med ; 176: 108530, 2024 Jun.
Article En | MEDLINE | ID: mdl-38749324

As an autoimmune-mediated inflammatory demyelinating disease of the central nervous system, multiple sclerosis (MS) is often confused with cerebral small vessel disease (cSVD), which is a regional pathological change in brain tissue with unknown pathogenesis. This is due to their similar clinical presentations and imaging manifestations. That misdiagnosis can significantly increase the occurrence of adverse events. Delayed or incorrect treatment is one of the most important causes of MS progression. Therefore, the development of a practical diagnostic imaging aid could significantly reduce the risk of misdiagnosis and improve patient prognosis. We propose an interpretable deep learning (DL) model that differentiates MS and cSVD using T2-weighted fluid-attenuated inversion recovery (FLAIR) images. Transfer learning (TL) was utilized to extract features from the ImageNet dataset. This pioneering model marks the first of its kind in neuroimaging, showing great potential in enhancing differential diagnostic capabilities within the field of neurological disorders. Our model extracts the texture features of the images and achieves more robust feature learning through two attention modules. The attention maps provided by the attention modules provide model interpretation to validate model learning and reveal more information to physicians. Finally, the proposed model is trained end-to-end using focal loss to reduce the influence of class imbalance. The model was validated using clinically diagnosed MS (n=112) and cSVD (n=321) patients from the Beijing Tiantan Hospital. The performance of the proposed model was better than that of two commonly used DL approaches, with a mean balanced accuracy of 86.06 % and a mean area under the receiver operating characteristic curve of 98.78 %. Moreover, the generated attention heat maps showed that the proposed model could focus on the lesion signatures in the image. The proposed model provides a practical diagnostic imaging aid for the use of routinely available imaging techniques such as magnetic resonance imaging to classify MS and cSVD by linking DL to human brain disease. We anticipate a substantial improvement in accurately distinguishing between various neurological conditions through this novel model.


Cerebral Small Vessel Diseases , Deep Learning , Multiple Sclerosis , Humans , Cerebral Small Vessel Diseases/diagnostic imaging , Multiple Sclerosis/diagnostic imaging , Male , Magnetic Resonance Imaging/methods , Female , Neural Networks, Computer , Image Interpretation, Computer-Assisted/methods , Middle Aged , Adult , Neuroimaging/methods
13.
Comput Biol Med ; 176: 108572, 2024 Jun.
Article En | MEDLINE | ID: mdl-38749327

BACKGROUND AND OBJECTIVE: Melanoma, a malignant form of skin cancer, is a critical health concern worldwide. Early and accurate detection plays a pivotal role in improving patient's conditions. Current diagnosis of skin cancer largely relies on visual inspections such as dermoscopy examinations, clinical screening and histopathological examinations. However, these approaches are characterized by low efficiency, high costs, and a lack of guaranteed accuracy. Consequently, deep learning based techniques have emerged in the field of melanoma detection, successfully aiding in improving the accuracy of diagnosis. However, the high similarity between benign and malignant melanomas, combined with the class imbalance issue in skin lesion datasets, present a significant challenge in further improving the diagnosis accuracy. We propose a two-stage framework for melanoma detection to address these issues. METHODS: In the first stage, we use Style Generative Adversarial Networks with Adaptive discriminator augmentation synthesis to generate realistic and diverse melanoma images, which are then combined with the original dataset to create an augmented dataset. In the second stage, we utilize a vision Transformer of BatchFormer to extract features and detect melanoma or non-melanoma skin lesions on the augmented dataset obtained in the previous step, specifically, we employed a dual-branch training strategy in this process. RESULTS: Our experimental results on the ISIC2020 dataset demonstrate the effectiveness of the proposed approach, showing a significant improvement in melanoma detection. The method achieved an accuracy of 98.43%, an AUC value of 98.63%, and an F1 value of 99.01%, surpassing some existing methods. CONCLUSION: The method is feasible, efficient, and achieves early melanoma screening. It significantly enhances detection accuracy and can assist physicians in diagnosis to a great extent.


Melanoma , Skin Neoplasms , Melanoma/diagnostic imaging , Melanoma/diagnosis , Humans , Skin Neoplasms/diagnostic imaging , Skin Neoplasms/diagnosis , Image Interpretation, Computer-Assisted/methods , Deep Learning , Dermoscopy/methods
14.
Comput Methods Programs Biomed ; 251: 108229, 2024 Jun.
Article En | MEDLINE | ID: mdl-38761413

BACKGROUND AND OBJECTIVE: Optical coherence tomography (OCT) is currently one of the most advanced retinal imaging methods. Retinal biomarkers in OCT images are of clinical significance and can assist ophthalmologists in diagnosing lesions. Compared with fundus images, OCT can provide higher resolution segmentation. However, image annotation at the bounding box level needs to be performed by ophthalmologists carefully and is difficult to obtain. In addition, the large variation in shape of different retinal markers and the inconspicuous appearance of biomarkers make it difficult for existing deep learning-based methods to effectively detect them. To overcome the above challenges, we propose a novel network for the detection of retinal biomarkers in OCT images. METHODS: We first address the issue of labeling cost using a novel weakly semi-supervised object detection method with point annotations which can reduce bounding box-level annotation efforts. To extend the method to the detection of biomarkers in OCT images, we propose multiple consistent regularizations for point-to-box regression network to deal with the shortage of supervision, which aims to learn more accurate regression mappings. Furthermore, in the subsequent fully supervised detection, we propose a cross-scale feature enhancement module to alleviate the detection problems caused by the large-scale variation of biomarkers. We also propose a dynamic label assignment strategy to distinguish samples of different importance more flexibly, thereby reducing detection errors due to the indistinguishable appearance of the biomarkers. RESULTS: When using our detection network, our regressor also achieves an AP value of 20.83 s when utilizing a 5 % fully labeled dataset partition, surpassing the performance of other comparative methods at 5 % and 10 %. Even coming close to the 20.87 % result achieved by Point DETR under 20 % full labeling conditions. When using Group R-CNN as the point-to-box regressor, our detector achieves 27.21 % AP in the 50 % fully labeled dataset experiment. 7.42 % AP improvement is achieved compared to our detection network baseline Faster R-CNN. CONCLUSIONS: The experimental findings not only demonstrate the effectiveness of our approach with minimal bounding box annotations but also highlight the enhanced biomarker detection performance of the proposed module. We have included a detailed algorithmic flow in the supplementary material.


Algorithms , Biomarkers , Retina , Tomography, Optical Coherence , Tomography, Optical Coherence/methods , Humans , Retina/diagnostic imaging , Deep Learning , Image Processing, Computer-Assisted/methods , Supervised Machine Learning , Neural Networks, Computer , Image Interpretation, Computer-Assisted/methods
15.
Comput Biol Med ; 176: 108585, 2024 Jun.
Article En | MEDLINE | ID: mdl-38761499

Active learning (AL) attempts to select informative samples in a dataset to minimize the number of required labels while maximizing the performance of the model. Current AL in segmentation tasks is limited to the expansion of popular classification-based methods including entropy, MC-dropout, etc. Meanwhile, most applications in the medical field are simply migrations that fail to consider the nature of medical images, such as high class imbalance, high domain difference, and data scarcity. In this study, we address these challenges and propose a novel AL framework for medical image segmentation task. Our approach introduces a pseudo-label-based filter addressing excessive blank patches in medical abnormalities segmentation tasks, e.g., lesions, and tumors, used before the AL selection. This filter helps reduce resource usage and allows the model to focus on selecting more informative samples. For the sample selection, we propose a novel query strategy that combines both model impact and data stability by employing adversarial attack. Furthermore, we harness the adversarial samples generated during the query process to enhance the robustness of the model. The experimental results verify our framework's effectiveness over various state-of-the-art methods. Our proposed method only needs less than 14% annotated patches in 3D brain MRI multiple sclerosis (MS) segmentation tasks and 20% for Low-Grade Glioma (LGG) tumor segmentation to achieve competitive results with full supervision. These promising outcomes not only improve performance but alleviate the time burden associated with expert annotation, thereby facilitating further advancements in the field of medical image segmentation. Our code is available at https://github.com/HelenMa9998/adversarial_active_learning.


Brain Neoplasms , Humans , Brain Neoplasms/diagnostic imaging , Magnetic Resonance Imaging/methods , Image Interpretation, Computer-Assisted/methods
16.
Comput Biol Med ; 176: 108594, 2024 Jun.
Article En | MEDLINE | ID: mdl-38761501

Skin cancer is one of the common types of cancer. It spreads quickly and is not easy to detect in the early stages, posing a major threat to human health. In recent years, deep learning methods have attracted widespread attention for skin cancer detection in dermoscopic images. However, training a practical classifier becomes highly challenging due to inter-class similarity and intra-class variation in skin lesion images. To address these problems, we propose a multi-scale fusion structure that combines shallow and deep features for more accurate classification. Simultaneously, we implement three approaches to the problem of class imbalance: class weighting, label smoothing, and resampling. In addition, the HAM10000_RE dataset strips out hair features to demonstrate the role of hair features in the classification process. We demonstrate that the region of interest is the most critical classification feature for the HAM10000_SE dataset, which segments lesion regions. We evaluated the effectiveness of our model using the HAM10000 and ISIC2019 dataset. The results showed that this method performed well in dermoscopic classification tasks, with ACC and AUC of 94.0% and 99.3%, on the HAM10000 dataset and ACC of 89.8% for the ISIC2019 dataset. The overall performance of our model is excellent in comparison to state-of-the-art models.


Dermoscopy , Skin Neoplasms , Humans , Skin Neoplasms/diagnostic imaging , Skin Neoplasms/pathology , Skin Neoplasms/classification , Dermoscopy/methods , Deep Learning , Image Interpretation, Computer-Assisted/methods , Skin/diagnostic imaging , Skin/pathology , Databases, Factual , Algorithms
17.
Comput Biol Med ; 176: 108590, 2024 Jun.
Article En | MEDLINE | ID: mdl-38763066

Over the past two decades, machine analysis of medical imaging has advanced rapidly, opening up significant potential for several important medical applications. As complicated diseases increase and the number of cases rises, the role of machine-based imaging analysis has become indispensable. It serves as both a tool and an assistant to medical experts, providing valuable insights and guidance. A particularly challenging task in this area is lesion segmentation, a task that is challenging even for experienced radiologists. The complexity of this task highlights the urgent need for robust machine learning approaches to support medical staff. In response, we present our novel solution: the D-TrAttUnet architecture. This framework is based on the observation that different diseases often target specific organs. Our architecture includes an encoder-decoder structure with a composite Transformer-CNN encoder and dual decoders. The encoder includes two paths: the Transformer path and the Encoders Fusion Module path. The Dual-Decoder configuration uses two identical decoders, each with attention gates. This allows the model to simultaneously segment lesions and organs and integrate their segmentation losses. To validate our approach, we performed evaluations on the Covid-19 and Bone Metastasis segmentation tasks. We also investigated the adaptability of the model by testing it without the second decoder in the segmentation of glands and nuclei. The results confirmed the superiority of our approach, especially in Covid-19 infections and the segmentation of bone metastases. In addition, the hybrid encoder showed exceptional performance in the segmentation of glands and nuclei, solidifying its role in modern medical image analysis.


Neural Networks, Computer , Humans , COVID-19/diagnostic imaging , SARS-CoV-2 , Machine Learning , Image Processing, Computer-Assisted/methods , Image Interpretation, Computer-Assisted/methods
18.
Curr Med Imaging ; 20(1): e15734056269264, 2024.
Article En | MEDLINE | ID: mdl-38766836

BACKGROUND: Currently, it is difficult to find a solution to the inverse inappropriate problem, which involves restoring a high-resolution image from a lowresolution image contained within a single image. In nature photography, one can capture a wide variety of objects and textures, each with its own characteristics, most notably the high-frequency component. These qualities can be distinguished from each other by looking at the pictures. OBJECTIVE: The goal is to develop an automated approach to identify thyroid nodules on ultrasound images. The aim of this research is to accurately differentiate thyroid nodules using Deep Learning Technique and to evaluate the effectiveness of different localization techniques. METHODS: The method used in this research is to reconstruct a single super-resolution image based on segmentation and classification. The poor-quality ultrasound image is divided into several parts, and the best applicable classification is chosen for each component. Pairs of high- and lowresolution images belonging to the same class are found and used to figure out which image is high-resolution for each segment. Deep learning technology, specifically the Adam classifier, is used to identify carcinoid tumors within thyroid nodules. Measures, such as localization accuracy, sensitivity, specificity, dice loss, ROC, and area under the curve (AUC), are used to evaluate the effectiveness of the techniques. RESULTS: The results of the proposed method are superior, both statistically and qualitatively, compared to other methods that are considered one of the latest and best technologies. The developed automated approach shows promising results in accurately identifying thyroid nodules on ultrasound images. CONCLUSION: The research demonstrates the development of an automated approach to identify thyroid nodules within ultrasound images using super-resolution single-image reconstruction and deep learning technology. The results indicate that the proposed method is superior to the latest and best techniques in terms of accuracy and quality. This research contributes to the advancement of medical imaging and holds the potential to improve the diagnosis and treatment of thyroid nodules.

.


Deep Learning , Thyroid Nodule , Ultrasonography , Humans , Thyroid Nodule/diagnostic imaging , Ultrasonography/methods , Thyroid Gland/diagnostic imaging , Image Interpretation, Computer-Assisted/methods , Image Processing, Computer-Assisted/methods
19.
Clin Radiol ; 79(7): e892-e899, 2024 Jul.
Article En | MEDLINE | ID: mdl-38719689

PURPOSE: We aimed to evaluate the feasibility of non-contrast-enhanced T1 sequence in texture analysis of breast cancer lesions to predict their estrogen receptor status. METHODS: The study included 85 pathologically proven breast cancer lesions in 53 patients. Immunohistochemical studies were performed to determine the estrogen receptor status (ER). Lesions were divided into two groups: ER + ve status and ER-ve status. Texture analysis using the second-order analysis features [The Co-occurrence matrix (11 features)] was applied on both T1 and dynamic contrast-enhanced (DCE) MRI images for each lesion. Texture features gained from both T1 and DCE images were analyzed to obtain cut-off values using ROC curves to sort lesions according to their estrogen receptor status. RESULTS: Angular second momentum and some of the entropy-based features showed statistically significant cut-off values in differentiation between the two groups [P-values for pre- and post-contrast images for AngSecMom (0.001, 0.008), sum entropy (0.003,0.005), and entropy (0.033,0.019), respectively]. On comparing the AUCs between pre- and post-contrast images, we found that differences were statistically insignificant. Sum of squares, sum variance, and sum average showed statistically significant cut-off points only on pre-contrast images [P-values for sum of squares (0.018), sum variance (0.024), and sum average (0.039)]. CONCLUSIONS: Texture analysis features showed promising results in predicting estrogen receptor status of breast cancer lesions on non-contrast T1 images.


Breast Neoplasms , Feasibility Studies , Magnetic Resonance Imaging , Receptors, Estrogen , Humans , Breast Neoplasms/diagnostic imaging , Breast Neoplasms/metabolism , Female , Magnetic Resonance Imaging/methods , Middle Aged , Receptors, Estrogen/metabolism , Adult , Aged , Image Interpretation, Computer-Assisted/methods , Breast/diagnostic imaging , Contrast Media , Retrospective Studies
20.
PLoS One ; 19(5): e0303670, 2024.
Article En | MEDLINE | ID: mdl-38820462

Breast cancer remains a critical global concern, underscoring the urgent need for early detection and accurate diagnosis to improve survival rates among women. Recent developments in deep learning have shown promising potential for computer-aided detection (CAD) systems to address this challenge. In this study, a novel segmentation method based on deep learning is designed to detect tumors in breast ultrasound images. Our proposed approach combines two powerful attention mechanisms: the novel Positional Convolutional Block Attention Module (PCBAM) and Shifted Window Attention (SWA), integrated into a Residual U-Net model. The PCBAM enhances the Convolutional Block Attention Module (CBAM) by incorporating the Positional Attention Module (PAM), thereby improving the contextual information captured by CBAM and enhancing the model's ability to capture spatial relationships within local features. Additionally, we employ SWA within the bottleneck layer of the Residual U-Net to further enhance the model's performance. To evaluate our approach, we perform experiments using two widely used datasets of breast ultrasound images and the obtained results demonstrate its capability in accurately detecting tumors. Our approach achieves state-of-the-art performance with dice score of 74.23% and 78.58% on BUSI and UDIAT datasets, respectively in segmenting the breast tumor region, showcasing its potential to help with precise tumor detection. By leveraging the power of deep learning and integrating innovative attention mechanisms, our study contributes to the ongoing efforts to improve breast cancer detection and ultimately enhance women's survival rates. The source code of our work can be found here: https://github.com/AyushRoy2001/DAUNet.


Breast Neoplasms , Deep Learning , Ultrasonography, Mammary , Humans , Breast Neoplasms/diagnostic imaging , Female , Ultrasonography, Mammary/methods , Neural Networks, Computer , Algorithms , Image Interpretation, Computer-Assisted/methods , Breast/diagnostic imaging , Breast/pathology , Image Processing, Computer-Assisted/methods
...