Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 141
Filtrar
1.
BMC Med Imaging ; 24(1): 63, 2024 Mar 18.
Artigo em Inglês | MEDLINE | ID: mdl-38500083

RESUMO

Significant advancements in machine learning algorithms have the potential to aid in the early detection and prevention of cancer, a devastating disease. However, traditional research methods face obstacles, and the amount of cancer-related information is rapidly expanding. The authors have developed a helpful support system using three distinct deep-learning models, ResNet-50, EfficientNet-B3, and ResNet-101, along with transfer learning, to predict lung cancer, thereby contributing to health and reducing the mortality rate associated with this condition. This offer aims to address the issue effectively. Using a dataset of 1,000 DICOM lung cancer images from the LIDC-IDRI repository, each image is classified into four different categories. Although deep learning is still making progress in its ability to analyze and understand cancer data, this research marks a significant step forward in the fight against cancer, promoting better health outcomes and potentially lowering the mortality rate. The Fusion Model, like all other models, achieved 100% precision in classifying Squamous Cells. The Fusion Model and ResNet-50 achieved a precision of 90%, closely followed by EfficientNet-B3 and ResNet-101 with slightly lower precision. To prevent overfitting and improve data collection and planning, the authors implemented a data extension strategy. The relationship between acquiring knowledge and reaching specific scores was also connected to advancing and addressing the issue of imprecise accuracy, ultimately contributing to advancements in health and a reduction in the mortality rate associated with lung cancer.


Assuntos
Aprendizado Profundo , Neoplasias Pulmonares , Humanos , Neoplasias Pulmonares/diagnóstico por imagem , Algoritmos , Aprendizado de Máquina , Projetos de Pesquisa
2.
BMC Med Imaging ; 24(1): 107, 2024 May 11.
Artigo em Inglês | MEDLINE | ID: mdl-38734629

RESUMO

This study addresses the critical challenge of detecting brain tumors using MRI images, a pivotal task in medical diagnostics that demands high accuracy and interpretability. While deep learning has shown remarkable success in medical image analysis, there remains a substantial need for models that are not only accurate but also interpretable to healthcare professionals. The existing methodologies, predominantly deep learning-based, often act as black boxes, providing little insight into their decision-making process. This research introduces an integrated approach using ResNet50, a deep learning model, combined with Gradient-weighted Class Activation Mapping (Grad-CAM) to offer a transparent and explainable framework for brain tumor detection. We employed a dataset of MRI images, enhanced through data augmentation, to train and validate our model. The results demonstrate a significant improvement in model performance, with a testing accuracy of 98.52% and precision-recall metrics exceeding 98%, showcasing the model's effectiveness in distinguishing tumor presence. The application of Grad-CAM provides insightful visual explanations, illustrating the model's focus areas in making predictions. This fusion of high accuracy and explainability holds profound implications for medical diagnostics, offering a pathway towards more reliable and interpretable brain tumor detection tools.


Assuntos
Neoplasias Encefálicas , Aprendizado Profundo , Imageamento por Ressonância Magnética , Humanos , Neoplasias Encefálicas/diagnóstico por imagem , Imageamento por Ressonância Magnética/métodos , Interpretação de Imagem Assistida por Computador/métodos
3.
BMC Med Imaging ; 24(1): 176, 2024 Jul 19.
Artigo em Inglês | MEDLINE | ID: mdl-39030496

RESUMO

Medical imaging stands as a critical component in diagnosing various diseases, where traditional methods often rely on manual interpretation and conventional machine learning techniques. These approaches, while effective, come with inherent limitations such as subjectivity in interpretation and constraints in handling complex image features. This research paper proposes an integrated deep learning approach utilizing pre-trained models-VGG16, ResNet50, and InceptionV3-combined within a unified framework to improve diagnostic accuracy in medical imaging. The method focuses on lung cancer detection using images resized and converted to a uniform format to optimize performance and ensure consistency across datasets. Our proposed model leverages the strengths of each pre-trained network, achieving a high degree of feature extraction and robustness by freezing the early convolutional layers and fine-tuning the deeper layers. Additionally, techniques like SMOTE and Gaussian Blur are applied to address class imbalance, enhancing model training on underrepresented classes. The model's performance was validated on the IQ-OTH/NCCD lung cancer dataset, which was collected from the Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases over a period of three months in fall 2019. The proposed model achieved an accuracy of 98.18%, with precision and recall rates notably high across all classes. This improvement highlights the potential of integrated deep learning systems in medical diagnostics, providing a more accurate, reliable, and efficient means of disease detection.


Assuntos
Aprendizado Profundo , Neoplasias Pulmonares , Humanos , Neoplasias Pulmonares/diagnóstico por imagem , Tomografia Computadorizada por Raios X/métodos , Redes Neurais de Computação
4.
BMC Musculoskelet Disord ; 25(1): 250, 2024 Apr 01.
Artigo em Inglês | MEDLINE | ID: mdl-38561697

RESUMO

BACKGROUND: Ankle fractures are prevalent injuries that necessitate precise diagnostic tools. Traditional diagnostic methods have limitations that can be addressed using machine learning techniques, with the potential to improve accuracy and expedite diagnoses. METHODS: We trained various deep learning architectures, notably the Adapted ResNet50 with SENet capabilities, to identify ankle fractures using a curated dataset of radiographic images. Model performance was evaluated using common metrics like accuracy, precision, and recall. Additionally, Grad-CAM visualizations were employed to interpret model decisions. RESULTS: The Adapted ResNet50 with SENet capabilities consistently outperformed other models, achieving an accuracy of 93%, AUC of 95%, and recall of 92%. Grad-CAM visualizations provided insights into areas of the radiographs that the model deemed significant in its decisions. CONCLUSIONS: The Adapted ResNet50 model enhanced with SENet capabilities demonstrated superior performance in detecting ankle fractures, offering a promising tool to complement traditional diagnostic methods. However, continuous refinement and expert validation are essential to ensure optimal application in clinical settings.


Assuntos
Fraturas do Tornozelo , Humanos , Fraturas do Tornozelo/diagnóstico por imagem , Benchmarking , Aprendizado de Máquina
5.
BMC Med Inform Decis Mak ; 24(1): 23, 2024 Jan 24.
Artigo em Inglês | MEDLINE | ID: mdl-38267994

RESUMO

Prostate cancer, the most common cancer in men, is influenced by age, family history, genetics, and lifestyle factors. Early detection of prostate cancer using screening methods improves outcomes, but the balance between overdiagnosis and early detection remains debated. Using Deep Learning (DL) algorithms for prostate cancer detection offers a promising solution for accurate and efficient diagnosis, particularly in cases where prostate imaging is challenging. In this paper, we propose a Prostate Cancer Detection Model (PCDM) model for the automatic diagnosis of prostate cancer. It proves its clinical applicability to aid in the early detection and management of prostate cancer in real-world healthcare environments. The PCDM model is a modified ResNet50-based architecture that integrates faster R-CNN and dual optimizers to improve the performance of the detection process. The model is trained on a large dataset of annotated medical images, and the experimental results show that the proposed model outperforms both ResNet50 and VGG19 architectures. Specifically, the proposed model achieves high sensitivity, specificity, precision, and accuracy rates of 97.40%, 97.09%, 97.56%, and 95.24%, respectively.


Assuntos
Aprendizado Profundo , Neoplasias da Próstata , Masculino , Humanos , Próstata , Neoplasias da Próstata/diagnóstico por imagem , Algoritmos , Instalações de Saúde
6.
J Appl Clin Med Phys ; 25(1): e14210, 2024 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-37991141

RESUMO

OBJECTIVE: This study aims to develop a ResNet50-based deep learning model for focal liver lesion (FLL) classification in ultrasound images, comparing its performance with other models and prior research. METHODOLOGY: We retrospectively collected 581 ultrasound images from the Chulabhorn Hospital's HCC surveillance and screening project (2010-2018). The dataset comprised five classes: non-FLL, hepatic cyst (Cyst), hemangioma (HMG), focal fatty sparing (FFS), and hepatocellular carcinoma (HCC). We conducted 5-fold cross-validation after random dataset partitioning, enhancing training data with data augmentation. Our models used modified pre-trained ResNet50, GGN, ResNet18, and VGG16 architectures. Model performance, assessed via confusion matrices for sensitivity, specificity, and accuracy, was compared across models and with prior studies. RESULTS: ResNet50 outperformed other models, achieving a 5-fold cross-validation accuracy of 87 ± 2.2%. While VGG16 showed similar performance, it exhibited higher uncertainty. In the testing phase, the pretrained ResNet50 excelled in classifying non-FLL, cysts, and FFS. To compare with other research, ResNet50 surpassed the prior methods like two-layered feed-forward neural networks (FFNN) and CNN+ReLU in FLL diagnosis. CONCLUSION: ResNet50 exhibited good performance in FLL diagnosis, especially for HCC classification, suggesting its potential for developing computer-aided FLL diagnosis. However, further refinement is required for HCC and HMG classification in future studies.


Assuntos
Carcinoma Hepatocelular , Cistos , Neoplasias Hepáticas , Humanos , Carcinoma Hepatocelular/diagnóstico por imagem , Neoplasias Hepáticas/diagnóstico por imagem , Estudos Retrospectivos , Redes Neurais de Computação
7.
Sensors (Basel) ; 24(3)2024 Jan 29.
Artigo em Inglês | MEDLINE | ID: mdl-38339598

RESUMO

Surface reconstruction using neural networks has proven effective in reconstructing dense 3D surfaces through image-based neural rendering. Nevertheless, current methods are challenging when dealing with the intricate details of large-scale scenes. The high-fidelity reconstruction performance of neural rendering is constrained by the view sparsity and structural complexity of such scenes. In this paper, we present Res-NeuS, a method combining ResNet-50 and neural surface rendering for dense 3D reconstruction. Specifically, we present appearance embeddings: ResNet-50 is used to extract the appearance depth features of an image to further capture more scene details. We interpolate points near the surface and optimize their weights for the accurate localization of 3D surfaces. We introduce photometric consistency and geometric constraints to optimize 3D surfaces and eliminate geometric ambiguity existing in current methods. Finally, we design a 3D geometry automatic sampling to filter out uninteresting areas and reconstruct complex surface details in a coarse-to-fine manner. Comprehensive experiments demonstrate Res-NeuS's superior capability in the reconstruction of 3D surfaces in complex, large-scale scenes, and the harmful distance of the reconstructed 3D model is 0.4 times that of general neural rendering 3D reconstruction methods and 0.6 times that of traditional 3D reconstruction methods.

8.
Sensors (Basel) ; 24(6)2024 Mar 07.
Artigo em Inglês | MEDLINE | ID: mdl-38543988

RESUMO

Concrete is extensively used in the construction of infrastructure such as houses and bridges. However, the appearance of cracks in concrete structures over time can diminish their sealing and load-bearing capability, potentially leading to structural failures and disasters. The timely detection of cracks allows for repairs without the need to replace the entire structure, resulting in cost savings. Currently, manual inspection remains the predominant method for identifying concrete cracks. However, in today's increasingly complex construction environments, subjective errors may arise due to human vision and perception. The purpose of this work is to investigate and design an autonomous convolutional neural network-based concrete detection system that can identify cracks automatically and use that information to calculate the crack proportion. The experiment's findings show that the trained model can classify concrete cracks with an accuracy of 99.9%. Moreover, the clustering technique applied to crack images enables the clear identification of the percentage of cracks, which facilitates the development of concrete damage level detection over time.

9.
Sensors (Basel) ; 24(5)2024 Feb 26.
Artigo em Inglês | MEDLINE | ID: mdl-38475050

RESUMO

Latent Low-Rank Representation (LatLRR) has emerged as a prominent approach for fusing visible and infrared images. In this approach, images are decomposed into three fundamental components: the base part, salient part, and sparse part. The aim is to blend the base and salient features to reconstruct images accurately. However, existing methods often focus more on combining the base and salient parts, neglecting the importance of the sparse component, whereas we advocate for the comprehensive inclusion of all three parts generated from LatLRR image decomposition into the image fusion process, a novel proposition introduced in this study. Moreover, the effective integration of Convolutional Neural Network (CNN) technology with LatLRR remains challenging, particularly after the inclusion of sparse parts. This study utilizes fusion strategies involving weighted average, summation, VGG19, and ResNet50 in various combinations to analyze the fusion performance following the introduction of sparse parts. The research findings show a significant enhancement in fusion performance achieved through the inclusion of sparse parts in the fusion process. The suggested fusion strategy involves employing deep learning techniques for fusing both base parts and sparse parts while utilizing a summation strategy for the fusion of salient parts. The findings improve the performance of LatLRR-based methods and offer valuable insights for enhancement, leading to advancements in the field of image fusion.

10.
J Xray Sci Technol ; 32(1): 53-68, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38189730

RESUMO

BACKGROUND: With the rapid growth of Deep Neural Networks (DNN) and Computer-Aided Diagnosis (CAD), more significant works have been analysed for cancer related diseases. Skin cancer is the most hazardous type of cancer that cannot be diagnosed in the early stages. OBJECTIVE: The diagnosis of skin cancer is becoming a challenge to dermatologists as an abnormal lesion looks like an ordinary nevus at the initial stages. Therefore, early identification of lesions (origin of skin cancer) is essential and helpful for treating skin cancer patients effectively. The enormous development of automated skin cancer diagnosis systems significantly supports dermatologists. METHODS: This paper performs a classification of skin cancer by utilising various deep-learning frameworks after resolving the class Imbalance problem in the ISIC-2019 dataset. A fine-tuned ResNet-50 model is used to evaluate the performance of original data, augmented data, and after by adding the focal loss. Focal loss is the best technique to solve overfitting problems by assigning weights to hard misclassified images. RESULTS: Finally, augmented data with focal loss is given a good classification performance with 98.85% accuracy, 95.52% precision, and 95.93% recall. Matthews Correlation coefficient (MCC) is the best metric to evaluate the quality of multi-class images. It has given outstanding performance by using augmented data and focal loss.


Assuntos
Aprendizado Profundo , Nevo , Neoplasias Cutâneas , Humanos , Neoplasias Cutâneas/diagnóstico por imagem , Neoplasias Cutâneas/patologia , Nevo/patologia , Redes Neurais de Computação , Diagnóstico por Computador/métodos
11.
BMC Bioinformatics ; 24(1): 157, 2023 Apr 19.
Artigo em Inglês | MEDLINE | ID: mdl-37076790

RESUMO

BACKGROUND: Diabetic retinopathy (DR) produces bleeding, exudation, and new blood vessel formation conditions. DR can damage the retinal blood vessels and cause vision loss or even blindness. If DR is detected early, ophthalmologists can use lasers to create tiny burns around the retinal tears to inhibit bleeding and prevent the formation of new blood vessels, in order to prevent deterioration of the disease. The rapid improvement of deep learning has made image recognition an effective technology; it can avoid misjudgments caused by different doctors' evaluations and help doctors to predict the condition quickly. The aim of this paper is to adopt visualization and preprocessing in the ResNet-50 model to improve module calibration, to enable the model to predict DR accurately. RESULTS: This study compared the performance of the proposed method with other common CNNs models (Xception, AlexNet, VggNet-s, VggNet-16 and ResNet-50). In examining said models, the results alluded to an over-fitting phenomenon, and the outcome of the work demonstrates that the performance of the revised ResNet-50 (Train accuracy: 0.8395 and Test accuracy: 0.7432) is better than other common CNNs (that is, the revised structure of ResNet-50 could avoid the overfitting problem, decease the loss value, and reduce the fluctuation problem). CONCLUSIONS: This study proposed two approaches to designing the DR grading system: a standard operation procedure (SOP) for preprocessing the fundus image, and a revised structure of ResNet-50, including an adaptive learning rating to adjust the weight of layers, regularization and change the structure of ResNet-50, which was selected for its suitable features. It is worth noting that the purpose of this study was not to design the most accurate DR screening network, but to demonstrate the effect of the SOP of DR and the visualization of the revised ResNet-50 model. The results provided an insight to revise the structure of CNNs using the visualization tool.


Assuntos
Diabetes Mellitus , Retinopatia Diabética , Humanos , Retinopatia Diabética/diagnóstico , Fundo de Olho , Programas de Rastreamento/métodos , Calibragem
12.
Skin Res Technol ; 29(11): e13524, 2023 Nov.
Artigo em Inglês | MEDLINE | ID: mdl-38009016

RESUMO

INTRODUCTION: Particularly within the Internet of Medical Things (IoMT) context, skin lesion analysis is critical for precise diagnosis. To improve the accuracy and efficiency of skin lesion analysis, CAD systems play a crucial role. To segment and classify skin lesions from dermoscopy images, this study focuses on using hybrid deep learning techniques. METHOD: This research uses a hybrid deep learning model that combines two cutting-edge approaches: Mask Region-based Convolutional Neural Network (MRCNN) for semantic segmentation and ResNet50 for lesion detection. To pinpoint the precise location of a skin lesion, the MRCNN is used for border delineation. We amass a huge, annotated collection of dermoscopy images for thorough model training. The hybrid deep learning model to capture subtle representations of the images is trained from start to finish using this dataset. RESULTS: The experimental results using dermoscopy images show that the suggested hybrid method outperforms the current state-of-the-art methods. The model's capacity to segment lesions into distinct groups is demonstrated by a segmentation accuracy measurement of 95.49 percent. In addition, the classification of skin lesions shows great accuracy and dependability, which is a notable advancement over traditional methods. The model is put through its paces on the ISIC 2020 Challenge dataset, scoring a perfect 96.75% accuracy. Compared to current best practices in IoMT, segmentation and classification models perform exceptionally well. CONCLUSION: In conclusion, this paper's hybrid deep learning strategy is highly effective in skin lesion segmentation and classification. The results show that the model has the potential to improve diagnostic accuracy in the setting of IoMT, and it outperforms the current gold standards. The excellent results obtained on the ISIC 2020 Challenge dataset further confirm the viability and superiority of the suggested methodology for skin lesion analysis.


Assuntos
Aprendizado Profundo , Melanoma , Dermatopatias , Neoplasias Cutâneas , Humanos , Neoplasias Cutâneas/diagnóstico por imagem , Neoplasias Cutâneas/patologia , Melanoma/patologia , Dermoscopia/métodos , Dermatopatias/diagnóstico por imagem , Internet
13.
Ophthalmic Res ; 66(1): 1278-1285, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-37778337

RESUMO

INTRODUCTION: Artificial intelligence has real potential for early identification of ocular diseases such as glaucoma. An important challenge is the requirement for large databases properly selected, which are not easily obtained. We used a relatively original strategy: a glaucoma recognition algorithm trained with fundus images from public databases and then tested and retrained with a carefully selected patient database. METHODS: The study's supervised deep learning method was an adapted version of the ResNet-50 architecture previously trained from 10,658 optic head images (glaucomatous or non-glaucomatous) from seven public databases. A total of 1,158 new images labeled by experts from 616 patients were added. The images were categorized after clinical examination including visual fields in 304 (26%) control images or those with ocular hypertension and 347 (30%) images with early, 290 (25%) with moderate, and 217 (19%) with advanced glaucoma. The initial algorithm was tested using 30% of the selected glaucoma database and then re-trained with 70% of this database and tested again. RESULTS: The results in the initial sample showed an area under the curve (AUC) of 76% for all images, and 66% for early, 82% for moderate, and 84% for advanced glaucoma. After retraining the algorithm, the respective AUC results were 82%, 72%, 89%, and 91%. CONCLUSION: Using combined data from public databases and data selected and labeled by experts facilitated improvement of the system's precision and identified interesting possibilities for obtaining tools for automatic screening of glaucomatous eyes more affordably.


Assuntos
Aprendizado Profundo , Glaucoma , Humanos , Inteligência Artificial , Glaucoma/diagnóstico , Fundo de Olho , Algoritmos
14.
BMC Med Inform Decis Mak ; 23(1): 232, 2023 10 19.
Artigo em Inglês | MEDLINE | ID: mdl-37858107

RESUMO

BACKGROUND: Cardiac arrhythmia is a cardiovascular disorder characterized by disturbances in the heartbeat caused by electrical conduction anomalies in cardiac muscle. Clinically, ECG machines are utilized to diagnose and monitor cardiac arrhythmia noninvasively. Since ECG signals are dynamic in nature and depict various complex information, visual assessment and analysis are time consuming and very difficult. Therefore, an automated system that can assist physicians in the easy detection of arrhythmia is needed. METHOD: The main objective of this study was to create an automated deep learning model capable of accurately classifying ECG signals into three categories: cardiac arrhythmia (ARR), congestive heart failure (CHF), and normal sinus rhythm (NSR). To achieve this, ECG data from the MIT-BIH and BIDMC databases available on PhysioNet were preprocessed and segmented before being utilized for deep learning model training. Pretrained models, ResNet 50 and AlexNet, were fine-tuned and configured to achieve optimal classification results. The main outcome measures for evaluating the performance of the model were F-measure, recall, precision, sensitivity, specificity, and accuracy, obtained from a multi-class confusion matrix. RESULT: The proposed deep learning model showed overall classification accuracy of 99.2%, average sensitivity of 99.2%, average specificity of 99.6%, average recall, precision and F- measure of 99.2% of test data. CONCLUSION: The proposed work introduced a robust approach for the classification of arrhythmias in comparison with the most recent state of the art and will reduce the diagnosis time and error that occurs in the visual investigation of ECG signals.


Assuntos
Doenças Cardiovasculares , Aprendizado Profundo , Humanos , Processamento de Sinais Assistido por Computador , Arritmias Cardíacas/diagnóstico , Eletrocardiografia/métodos , Algoritmos
15.
J Appl Clin Med Phys ; 24(3): e13875, 2023 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-36546583

RESUMO

In this study, we investigated 3D convolutional neural networks (CNNs) with input from radiographic and dosimetric datasets of primary lung tumors and surrounding lung volumes to predict the likelihood of radiation pneumonitis (RP). Pre-treatment, 3- and 6-month follow-up computed tomography (CT) and 3D dose datasets from one hundred and ninety-three NSCLC patients treated with stereotactic body radiotherapy (SBRT) were retrospectively collected and analyzed for this study. DenseNet-121 and ResNet-50 models were selected for this study as they are deep neural networks and have been proven to have high accuracy for complex image classification tasks. Both were modified with 3D convolution and max pooling layers to accept 3D datasets. We used a minority class oversampling approach and data augmentation to address the challenges of data imbalance and data scarcity. We built two sets of models for classification of three (No RP, Grade 1 RP, Grade 2 RP) and two (No RP, Yes RP) classes as outputs. The 3D DenseNet-121 models performed better (F1 score [0.81], AUC [0.91] [three class]; F1 score [0.77], AUC [0.84] [two class]) than the 3D ResNet-50 models (F1 score [0.54], AUC [0.72] [three-class]; F1 score [0.68], AUC [0.71] [two-class]) (p = 0.017 for three class predictions). We also attempted to identify salient regions within the input 3D image dataset via integrated gradient (IG) techniques to assess the relevance of the tumor surrounding volume for RP stratification. These techniques appeared to indicate the significance of the tumor and surrounding regions in the prediction of RP. Overall, 3D CNNs performed well to predict clinical RP in our cohort based on the provided image sets and radiotherapy dose information.


Assuntos
Carcinoma Pulmonar de Células não Pequenas , Neoplasias Pulmonares , Pneumonite por Radiação , Radiocirurgia , Humanos , Radiocirurgia/efeitos adversos , Pneumonite por Radiação/diagnóstico , Pneumonite por Radiação/etiologia , Pneumonite por Radiação/patologia , Estudos Retrospectivos , Carcinoma Pulmonar de Células não Pequenas/radioterapia , Carcinoma Pulmonar de Células não Pequenas/cirurgia , Neoplasias Pulmonares/radioterapia , Neoplasias Pulmonares/cirurgia , Neoplasias Pulmonares/patologia , Redes Neurais de Computação
16.
Sensors (Basel) ; 23(18)2023 Sep 19.
Artigo em Inglês | MEDLINE | ID: mdl-37766026

RESUMO

Historically, individuals with hearing impairments have faced neglect, lacking the necessary tools to facilitate effective communication. However, advancements in modern technology have paved the way for the development of various tools and software aimed at improving the quality of life for hearing-disabled individuals. This research paper presents a comprehensive study employing five distinct deep learning models to recognize hand gestures for the American Sign Language (ASL) alphabet. The primary objective of this study was to leverage contemporary technology to bridge the communication gap between hearing-impaired individuals and individuals with no hearing impairment. The models utilized in this research include AlexNet, ConvNeXt, EfficientNet, ResNet-50, and VisionTransformer were trained and tested using an extensive dataset comprising over 87,000 images of the ASL alphabet hand gestures. Numerous experiments were conducted, involving modifications to the architectural design parameters of the models to obtain maximum recognition accuracy. The experimental results of our study revealed that ResNet-50 achieved an exceptional accuracy rate of 99.98%, the highest among all models. EfficientNet attained an accuracy rate of 99.95%, ConvNeXt achieved 99.51% accuracy, AlexNet attained 99.50% accuracy, while VisionTransformer yielded the lowest accuracy of 88.59%.


Assuntos
Aprendizado Profundo , Língua de Sinais , Humanos , Estados Unidos , Qualidade de Vida , Gestos , Tecnologia
17.
Sensors (Basel) ; 23(23)2023 Nov 30.
Artigo em Inglês | MEDLINE | ID: mdl-38067888

RESUMO

The primary objective of this study is to develop an advanced, automated system for the early detection and classification of leaf diseases in potato plants, which are among the most cultivated vegetable crops worldwide. These diseases, notably early and late blight caused by Alternaria solani and Phytophthora infestans, significantly impact the quantity and quality of global potato production. We hypothesize that the integration of Vision Transformer (ViT) and ResNet-50 architectures in a new model, named EfficientRMT-Net, can effectively and accurately identify various potato leaf diseases. This approach aims to overcome the limitations of traditional methods, which are often labor-intensive, time-consuming, and prone to inaccuracies due to the unpredictability of disease presentation. EfficientRMT-Net leverages the CNN model for distinct feature extraction and employs depth-wise convolution (DWC) to reduce computational demands. A stage block structure is also incorporated to improve scalability and sensitive area detection, enhancing transferability across different datasets. The classification tasks are performed using a global average pooling layer and a fully connected layer. The model was trained, validated, and tested on custom datasets specifically curated for potato leaf disease detection. EfficientRMT-Net's performance was compared with other deep learning and transfer learning techniques to establish its efficacy. Preliminary results show that EfficientRMT-Net achieves an accuracy of 97.65% on a general image dataset and 99.12% on a specialized Potato leaf image dataset, outperforming existing methods. The model demonstrates a high level of proficiency in correctly classifying and identifying potato leaf diseases, even in cases of distorted samples. The EfficientRMT-Net model provides an efficient and accurate solution for classifying potato plant leaf diseases, potentially enabling farmers to enhance crop yield while optimizing resource utilization. This study confirms our hypothesis, showcasing the effectiveness of combining ViT and ResNet-50 architectures in addressing complex agricultural challenges.


Assuntos
Solanum tuberosum , Agricultura , Produtos Agrícolas , Cultura , Doenças das Plantas , Folhas de Planta
18.
Sensors (Basel) ; 23(2)2023 Jan 04.
Artigo em Inglês | MEDLINE | ID: mdl-36679381

RESUMO

This article is devoted to the development of a classification method based on an artificial neural network architecture to solve the problem of recognizing the sources of acoustic influences recorded by a phase-sensitive OTDR. At the initial stage of signal processing, we propose the use of a band-pass filter to collect data sets with an increased signal-to-noise ratio. When solving the classification problem, we study three widely used convolutional neural network architectures: AlexNet, ResNet50, and DenseNet169. As a result of computational experiments, it is shown that the AlexNet and DenseNet169 architectures can obtain accuracies above 90%. In addition, we propose a novel CNN architecture based on AlexNet, which obtains the best results; in particular, its accuracy is above 98%. The advantages of the proposed model include low power consumption (400 mW) and high speed (0.032 s per net evaluation). In further studies, in order to increase the accuracy, reliability, and data invariance, the use of new algorithms for the filtering and extraction of acoustic signals recorded by a phase-sensitive reflectometer will be considered.


Assuntos
Algoritmos , Redes Neurais de Computação , Reprodutibilidade dos Testes , Razão Sinal-Ruído , Acústica
19.
Sensors (Basel) ; 23(15)2023 Jul 28.
Artigo em Inglês | MEDLINE | ID: mdl-37571532

RESUMO

Infrared thermography (IRT) is a technique used to diagnose Photovoltaic (PV) installations to detect sub-optimal conditions. The increase of PV installations in smart cities has generated the search for technology that improves the use of IRT, which requires irradiance conditions to be greater than 700 W/m2, making it impossible to use at times when irradiance goes under that value. This project presents an IoT platform working on artificial intelligence (AI) which automatically detects hot spots in PV modules by analyzing the temperature differentials between modules exposed to irradiances greater than 300 W/m2. For this purpose, two AI (Deep learning and machine learning) were trained and tested in a real PV installation where hot spots were induced. The system was able to detect hot spots with a sensitivity of 0.995 and an accuracy of 0.923 under dirty, short-circuited, and partially shaded conditions. This project differs from others because it proposes an alternative to facilitate the implementation of diagnostics with IRT and evaluates the real temperatures of PV modules, which represents a potential economic saving for PV installation managers and inspectors.

20.
Sensors (Basel) ; 23(17)2023 Aug 31.
Artigo em Inglês | MEDLINE | ID: mdl-37688036

RESUMO

Some recent studies show that filters in convolutional neural networks (CNNs) have low color selectivity in datasets of natural scenes such as Imagenet. CNNs, bio-inspired by the visual cortex, are characterized by their hierarchical learning structure which appears to gradually transform the representation space. Inspired by the direct connection between the LGN and V4, which allows V4 to handle low-level information closer to the trichromatic input in addition to processed information that comes from V2/V3, we propose the addition of a long skip connection (LSC) between the first and last blocks of the feature extraction stage to allow deeper parts of the network to receive information from shallower layers. This type of connection improves classification accuracy by combining simple-visual and complex-abstract features to create more color-selective ones. We have applied this strategy to classic CNN architectures and quantitatively and qualitatively analyzed the improvement in accuracy while focusing on color selectivity. The results show that, in general, skip connections improve accuracy, but LSC improves it even more and enhances the color selectivity of the original CNN architectures. As a side result, we propose a new color representation procedure for organizing and filtering feature maps, making their visualization more manageable for qualitative color selectivity analysis.

SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA