Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 9 de 9
Filtrar
Mais filtros

Base de dados
Tipo de documento
Intervalo de ano de publicação
1.
Sensors (Basel) ; 22(15)2022 Jul 24.
Artigo em Inglês | MEDLINE | ID: mdl-35898023

RESUMO

Acute lymphoblastic leukemia (ALL) is a deadly cancer characterized by aberrant accumulation of immature lymphocytes in the blood or bone marrow. Effective treatment of ALL is strongly associated with the early diagnosis of the disease. Current practice for initial ALL diagnosis is performed through manual evaluation of stained blood smear microscopy images, which is a time-consuming and error-prone process. Deep learning-based human-centric biomedical diagnosis has recently emerged as a powerful tool for assisting physicians in making medical decisions. Therefore, numerous computer-aided diagnostic systems have been developed to autonomously identify ALL in blood images. In this study, a new Bayesian-based optimized convolutional neural network (CNN) is introduced for the detection of ALL in microscopic smear images. To promote classification performance, the architecture of the proposed CNN and its hyperparameters are customized to input data through the Bayesian optimization approach. The Bayesian optimization technique adopts an informed iterative procedure to search the hyperparameter space for the optimal set of network hyperparameters that minimizes an objective error function. The proposed CNN is trained and validated using a hybrid dataset which is formed by integrating two public ALL datasets. Data augmentation has been adopted to further supplement the hybrid image set to boost classification performance. The Bayesian search-derived optimal CNN model recorded an improved performance of image-based ALL classification on test set. The findings of this study reveal the superiority of the proposed Bayesian-optimized CNN over other optimized deep learning ALL classification models.


Assuntos
Redes Neurais de Computação , Leucemia-Linfoma Linfoblástico de Células Precursoras , Teorema de Bayes , Humanos , Microscopia , Leucemia-Linfoma Linfoblástico de Células Precursoras/diagnóstico
2.
Sensors (Basel) ; 22(13)2022 Jun 30.
Artigo em Inglês | MEDLINE | ID: mdl-35808433

RESUMO

One of the most promising research areas in the healthcare industry and the scientific community is focusing on the AI-based applications for real medical challenges such as the building of computer-aided diagnosis (CAD) systems for breast cancer. Transfer learning is one of the recent emerging AI-based techniques that allow rapid learning progress and improve medical imaging diagnosis performance. Although deep learning classification for breast cancer has been widely covered, certain obstacles still remain to investigate the independency among the extracted high-level deep features. This work tackles two challenges that still exist when designing effective CAD systems for breast lesion classification from mammograms. The first challenge is to enrich the input information of the deep learning models by generating pseudo-colored images instead of only using the input original grayscale images. To achieve this goal two different image preprocessing techniques are parallel used: contrast-limited adaptive histogram equalization (CLAHE) and Pixel-wise intensity adjustment. The original image is preserved in the first channel, while the other two channels receive the processed images, respectively. The generated three-channel pseudo-colored images are fed directly into the input layer of the backbone CNNs to generate more powerful high-level deep features. The second challenge is to overcome the multicollinearity problem that occurs among the high correlated deep features generated from deep learning models. A new hybrid processing technique based on Logistic Regression (LR) as well as Principal Components Analysis (PCA) is presented and called LR-PCA. Such a process helps to select the significant principal components (PCs) to further use them for the classification purpose. The proposed CAD system has been examined using two different public benchmark datasets which are INbreast and mini-MAIS. The proposed CAD system could achieve the highest performance accuracies of 98.60% and 98.80% using INbreast and mini-MAIS datasets, respectively. Such a CAD system seems to be useful and reliable for breast cancer diagnosis.


Assuntos
Neoplasias da Mama , Redes Neurais de Computação , Neoplasias da Mama/diagnóstico por imagem , Feminino , Humanos , Modelos Logísticos , Aprendizado de Máquina , Mamografia/métodos
3.
Entropy (Basel) ; 23(10)2021 Sep 26.
Artigo em Inglês | MEDLINE | ID: mdl-34681974

RESUMO

Diabetic macular edema (DME) is the most common cause of irreversible vision loss in diabetes patients. Early diagnosis of DME is necessary for effective treatment of the disease. Visual detection of DME in retinal screening images by ophthalmologists is a time-consuming process. Recently, many computer-aided diagnosis systems have been developed to assist doctors by detecting DME automatically. In this paper, a new deep feature transfer-based stacked autoencoder neural network system is proposed for the automatic diagnosis of DME in fundus images. The proposed system integrates the power of pretrained convolutional neural networks as automatic feature extractors with the power of stacked autoencoders in feature selection and classification. Moreover, the system enables extracting a large set of features from a small input dataset using four standard pretrained deep networks: ResNet-50, SqueezeNet, Inception-v3, and GoogLeNet. The most informative features are then selected by a stacked autoencoder neural network. The stacked network is trained in a semi-supervised manner and is used for the classification of DME. It is found that the introduced system achieves a maximum classification accuracy of 96.8%, sensitivity of 97.5%, and specificity of 95.5%. The proposed system shows a superior performance over the original pretrained network classifiers and state-of-the-art findings.

4.
PeerJ Comput Sci ; 10: e1756, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38196952

RESUMO

The telecom sector is currently undergoing a digital transformation by integrating artificial intelligence (AI) and Internet of Things (IoT) technologies. Customer retention in this context relies on the application of autonomous AI methods for analyzing IoT device data patterns in relation to the offered service packages. One significant challenge in existing studies is treating churn recognition and customer segmentation as separate tasks, which diminishes overall system accuracy. This study introduces an innovative approach by leveraging a unified customer analytics platform that treats churn recognition and segmentation as a bi-level optimization problem. The proposed framework includes an Auto Machine Learning (AutoML) oversampling method, effectively handling three mixed datasets of customer churn features while addressing imbalanced-class distribution issues. To enhance performance, the study utilizes the strength of oversampling methods like synthetic minority oversampling technique for nominal and continuous features (SMOTE-NC) and synthetic minority oversampling with encoded nominal and continuous features (SMOTE-ENC). Performance evaluation, using 10-fold cross-validation, measures accuracy and F1-score. Simulation results demonstrate that the proposed strategy, particularly Random Forest (RF) with SMOTE-NC, outperforms standard methods with SMOTE. It achieves accuracy rates of 79.24%, 94.54%, and 69.57%, and F1-scores of 65.25%, 81.87%, and 45.62% for the IBM, Kaggle Telco and Cell2Cell datasets, respectively. The proposed method autonomously determines the number and density of clusters. Factor analysis employing Bayesian logistic regression identifies influential factors for accurate customer segmentation. Furthermore, the study segments consumers behaviorally and generates targeted recommendations for personalized service packages, benefiting decision-makers.

5.
Heliyon ; 10(7): e28967, 2024 Apr 15.
Artigo em Inglês | MEDLINE | ID: mdl-38601589

RESUMO

Plant diseases annually cause damage and loss of much of the crop, if not its complete destruction, and this constitutes a significant challenge for farm owners, governments, and consumers alike. Therefore, identifying and classifying diseases at an early stage is very important in order to sustain local and global food security. In this research, we designed a new method to identify plant diseases by combining transfer learning and Gravitational Search Algorithm (GSA). Two state-of-the-art pretrained models have been adopted for extracting features in this study, which are MobileNetV2 and ResNe50V2. Multilayer feature extraction is applied in this study to ensure representations of plant leaves from different levels of abstraction for precise classification. These features are then concatenated and passed to GSA for optimizing them. Finally, optimized features are passed to Multinomial Logistic Regression (MLR) for final classification. This integration is essential for categorizing 18 different types of infected and healthy leaf samples. The performance of our approach is strengthened by a comparative analysis that incorporates features optimized by the Genetic Algorithm (GA). Additionally, the MLR algorithm is contrasted with K-Nearest Neighbors (KNN). The empirical findings indicate that our model, which has been refined using GSA, achieves very high levels of precision. Specifically, the average precision for MLR is 99.2%, while for KNN it is 98.6%. The resulting results significantly exceed those achieved with GA-optimized features, thereby highlighting the superiority of our suggested strategy. One important result of our study is that we were able to decrease the number of features by more than 50%. This reduction greatly reduces the processing requirements without sacrificing the quality of the diagnosis. This work presents a robust and efficient approach to the early detection of plant diseases. The work demonstrates the utilization of sophisticated computational methods in agriculture, enabling the development of novel data-driven strategies for plant health management, therefore enhancing worldwide food security.

6.
Diagnostics (Basel) ; 13(16)2023 Aug 14.
Artigo em Inglês | MEDLINE | ID: mdl-37627931

RESUMO

Acute lymphoblastic leukemia (ALL) is a lethal blood cancer that is characterized by an abnormal increased number of immature lymphocytes in the blood or bone marrow. For effective treatment of ALL, early assessment of the disease is essential. Manual examination of stained blood smear images is current practice for initially screening ALL. This practice is time-consuming and error-prone. In order to effectively diagnose ALL, numerous deep-learning-based computer vision systems have been developed for detecting ALL in blood peripheral images (BPIs). Such systems extract a huge number of image features and use them to perform the classification task. The extracted features may contain irrelevant or redundant features that could reduce classification accuracy and increase the running time of the classifier. Feature selection is considered an effective tool to mitigate the curse of the dimensionality problem and alleviate its corresponding shortcomings. One of the most effective dimensionality-reduction tools is principal component analysis (PCA), which maps input features into an orthogonal space and extracts the features that convey the highest variability from the data. Other feature selection approaches utilize evolutionary computation (EC) to search the feature space and localize optimal features. To profit from both feature selection approaches in improving the classification performance of ALL, in this study, a new hybrid deep-learning-based feature engineering approach is proposed. The introduced approach integrates the powerful capability of PCA and particle swarm optimization (PSO) approaches in selecting informative features from BPI mages with the power of pre-trained CNNs of feature extraction. Image features are first extracted through the feature-transfer capability of the GoogleNet convolutional neural network (CNN). PCA is utilized to generate a feature set of the principal components that covers 95% of the variability in the data. In parallel, bio-inspired particle swarm optimization is used to search for the optimal image features. The PCA and PSO-derived feature sets are then integrated to develop a hybrid set of features that are then used to train a Bayesian-based optimized support vector machine (SVM) and subspace discriminant ensemble-learning (SDEL) classifiers. The obtained results show improved classification performance for the ML classifiers trained by the proposed hybrid feature set over the original PCA, PSO, and all extracted feature sets for ALL multi-class classification. The Bayesian-optimized SVM trained with the proposed hybrid PCA-PSO feature set achieves the highest classification accuracy of 97.4%. The classification performance of the proposed feature engineering approach competes with the state of the art.

7.
Healthcare (Basel) ; 11(21)2023 Oct 27.
Artigo em Inglês | MEDLINE | ID: mdl-37957985

RESUMO

In recent times, there has been considerable focus on harnessing artificial intelligence (AI) for medical image analysis and healthcare purposes. In this study, we introduce CADFU (Computer-Aided Diagnosis System for Foot Ulcers), a pioneering diabetic foot ulcer diagnosis system. The primary objective of CADFU is to detect and segment ulcers and similar chronic wounds in medical images. To achieve this, we employ two distinct algorithms. Firstly, DHuNeT, an innovative Dual-Phase Hyperactive UNet, is utilized for the segmentation task. Second, we used YOLOv8 for the task of detecting wounds. The DHuNeT autoencoder, employed for the wound segmentation task, is the paper's primary and most significant contribution. DHuNeT is the combination of sequentially stacking two UNet autoencoders. The hyperactive information transmission from the first UNet to the second UNet is the key idea of DHuNeT. The first UNet feeds the second UNet the features it has learned, and the two UNets combine their learned features to create new, more accurate, and effective features. We achieve good performance measures, especially in terms of the Dice co-efficient and precision, with segmentation scores of 85% and 92.6%, respectively. We obtain a mean average precision (mAP) of 86% in the detection task. Future hospitals could quickly monitor patients' health using the proposed CADFU system, which would be beneficial for both patients and doctors.

8.
Healthcare (Basel) ; 10(12)2022 Nov 22.
Artigo em Inglês | MEDLINE | ID: mdl-36553864

RESUMO

Brain tumors (BTs) are an uncommon but fatal kind of cancer. Therefore, the development of computer-aided diagnosis (CAD) systems for classifying brain tumors in magnetic resonance imaging (MRI) has been the subject of many research papers so far. However, research in this sector is still in its early stage. The ultimate goal of this research is to develop a lightweight effective implementation of the U-Net deep network for use in performing exact real-time segmentation. Moreover, a simplified deep convolutional neural network (DCNN) architecture for the BT classification is presented for automatic feature extraction and classification of the segmented regions of interest (ROIs). Five convolutional layers, rectified linear unit, normalization, and max-pooling layers make up the DCNN's proposed simplified architecture. The introduced method was verified on multimodal brain tumor segmentation (BRATS 2015) datasets. Our experimental results on BRATS 2015 acquired Dice similarity coefficient (DSC) scores, sensitivity, and classification accuracy of 88.8%, 89.4%, and 88.6% for high-grade gliomas. When it comes to segmenting BRATS 2015 BT images, the performance of our proposed CAD framework is on par with existing state-of-the-art methods. However, the accuracy achieved in this study for the classification of BT images has improved upon the accuracy reported in prior studies. Image classification accuracy for BRATS 2015 BT has been improved from 88% to 88.6%.

9.
Diagnostics (Basel) ; 12(10)2022 Oct 20.
Artigo em Inglês | MEDLINE | ID: mdl-36292230

RESUMO

Brain tumors (BTs) are deadly diseases that can strike people of every age, all over the world. Every year, thousands of people die of brain tumors. Brain-related diagnoses require caution, and even the smallest error in diagnosis can have negative repercussions. Medical errors in brain tumor diagnosis are common and frequently result in higher patient mortality rates. Magnetic resonance imaging (MRI) is widely used for tumor evaluation and detection. However, MRI generates large amounts of data, making manual segmentation difficult and laborious work, limiting the use of accurate measurements in clinical practice. As a result, automated and dependable segmentation methods are required. Automatic segmentation and early detection of brain tumors are difficult tasks in computer vision due to their high spatial and structural variability. Therefore, early diagnosis or detection and treatment are critical. Various traditional Machine learning (ML) techniques have been used to detect various types of brain tumors. The main issue with these models is that the features were manually extracted. To address the aforementioned insightful issues, this paper presents a hybrid deep transfer learning (GN-AlexNet) model of BT tri-classification (pituitary, meningioma, and glioma). The proposed model combines GoogleNet architecture with the AlexNet model by removing the five layers of GoogleNet and adding ten layers of the AlexNet model, which extracts features and classifies them automatically. On the same CE-MRI dataset, the proposed model was compared to transfer learning techniques (VGG-16, AlexNet, SqeezNet, ResNet, and MobileNet-V2) and ML/DL. The proposed model outperformed the current methods in terms of accuracy and sensitivity (accuracy of 99.51% and sensitivity of 98.90%).

SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA