Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 68
Filtrar
1.
Eur J Pediatr ; 2024 Jun 14.
Artigo em Inglês | MEDLINE | ID: mdl-38871980

RESUMO

Williams-Beuren syndrome (WBS) is a rare genetic disorder characterized by special facial gestalt, delayed development, and supravalvular aortic stenosis or/and stenosis of the branches of the pulmonary artery. We aim to develop and optimize accurate models of facial recognition to assist in the diagnosis of WBS, and to evaluate their effectiveness by using both five-fold cross-validation and an external test set. We used a total of 954 images from 135 patients with WBS, 124 patients suffering from other genetic disorders, and 183 healthy children. The training set comprised 852 images of 104 WBS cases, 91 cases of other genetic disorders, and 145 healthy children from September 2017 to December 2021 at the Guangdong Provincial People's Hospital. We constructed six binary classification models of facial recognition for WBS by using EfficientNet-b3, ResNet-50, VGG-16, VGG-16BN, VGG-19, and VGG-19BN. Transfer learning was used to pre-train the models, and each model was modified with a variable cosine learning rate. Each model was first evaluated by using five-fold cross-validation and then assessed on the external test set. The latter contained 102 images of 31 children suffering from WBS, 33 children with other genetic disorders, and 38 healthy children. To compare the capabilities of these models of recognition with those of human experts in terms of identifying cases of WBS, we recruited two pediatricians, a pediatric cardiologist, and a pediatric geneticist to identify the WBS patients based solely on their facial images. We constructed six models of facial recognition for diagnosing WBS using EfficientNet-b3, ResNet-50, VGG-16, VGG-16BN, VGG-19, and VGG-19BN. The model based on VGG-19BN achieved the best performance in terms of five-fold cross-validation, with an accuracy of 93.74% ± 3.18%, precision of 94.93% ± 4.53%, specificity of 96.10% ± 4.30%, and F1 score of 91.65% ± 4.28%, while the VGG-16BN model achieved the highest recall value of 91.63% ± 5.96%. The VGG-19BN model also achieved the best performance on the external test set, with an accuracy of 95.10%, precision of 100%, recall of 83.87%, specificity of 93.42%, and F1 score of 91.23%. The best performance by human experts on the external test set yielded values of accuracy, precision, recall, specificity, and F1 scores of 77.45%, 60.53%, 77.42%, 83.10%, and 66.67%, respectively. The F1 score of each human expert was lower than those of the EfficientNet-b3 (84.21%), ResNet-50 (74.51%), VGG-16 (85.71%), VGG-16BN (85.71%), VGG-19 (83.02%), and VGG-19BN (91.23%) models. CONCLUSION: The results showed that facial recognition technology can be used to accurately diagnose patients with WBS. Facial recognition models based on VGG-19BN can play a crucial role in its clinical diagnosis. Their performance can be improved by expanding the size of the training dataset, optimizing the CNN architectures applied, and modifying them with a variable cosine learning rate. WHAT IS KNOWN: • The facial gestalt of WBS, often described as "elfin," includes a broad forehead, periorbital puffiness, a flat nasal bridge, full cheeks, and a small chin. • Recent studies have demonstrated the potential of deep convolutional neural networks for facial recognition as a diagnostic tool for WBS. WHAT IS NEW: • This study develops six models of facial recognition, EfficientNet-b3, ResNet-50, VGG-16, VGG-16BN, VGG-19, and VGG-19BN, to improve WBS diagnosis. • The VGG-19BN model achieved the best performance, with an accuracy of 95.10% and specificity of 93.42%. The facial recognition model based on VGG-19BN can play a crucial role in the clinical diagnosis of WBS.

2.
BMC Med Imaging ; 24(1): 120, 2024 May 24.
Artigo em Inglês | MEDLINE | ID: mdl-38789925

RESUMO

BACKGROUND: Lung cancer is the second most common cancer worldwide, with over two million new cases per year. Early identification would allow healthcare practitioners to handle it more effectively. The advancement of computer-aided detection systems significantly impacted clinical analysis and decision-making on human disease. Towards this, machine learning and deep learning techniques are successfully being applied. Due to several advantages, transfer learning has become popular for disease detection based on image data. METHODS: In this work, we build a novel transfer learning model (VER-Net) by stacking three different transfer learning models to detect lung cancer using lung CT scan images. The model is trained to map the CT scan images with four lung cancer classes. Various measures, such as image preprocessing, data augmentation, and hyperparameter tuning, are taken to improve the efficacy of VER-Net. All the models are trained and evaluated using multiclass classifications chest CT images. RESULTS: The experimental results confirm that VER-Net outperformed the other eight transfer learning models compared with. VER-Net scored 91%, 92%, 91%, and 91.3% when tested for accuracy, precision, recall, and F1-score, respectively. Compared to the state-of-the-art, VER-Net has better accuracy. CONCLUSION: VER-Net is not only effectively used for lung cancer detection but may also be useful for other diseases for which CT scan images are available.


Assuntos
Neoplasias Pulmonares , Tomografia Computadorizada por Raios X , Humanos , Neoplasias Pulmonares/diagnóstico por imagem , Tomografia Computadorizada por Raios X/métodos , Aprendizado de Máquina , Aprendizado Profundo , Interpretação de Imagem Radiográfica Assistida por Computador/métodos
3.
Postgrad Med J ; 100(1186): 592-602, 2024 Jul 18.
Artigo em Inglês | MEDLINE | ID: mdl-38507237

RESUMO

PURPOSE: To construct a clinical noncontrastive computed tomography (NCCT) deep learning joint model for predicting early hematoma expansion (HE) after cerebral hemorrhage (sICH) and evaluate its predictive performance. METHODS: All 254 patients with primary cerebral hemorrhage from January 2017 to December 2022 in the General Hospital of the Western Theater Command were included. According to the criteria of hematoma enlargement exceeding 33% or the volume exceeding 6 ml, the patients were divided into the HE group and the hematoma non-enlargement (NHE) group. Multiple models and the 10-fold cross-validation method were used to screen the most valuable features and model the probability of predicting HE. The area under the curve (AUC) was used to analyze the prediction efficiency of each model for HE. RESULTS: They were randomly divided into a training set of 204 cases in an 8:2 ratio and 50 cases of the test set. The clinical imaging deep feature joint model (22 features) predicted the area under the curve of HE as follows: clinical Navie Bayes model AUC 0.779, traditional radiology logistic regression (LR) model AUC 0.818, deep learning LR model AUC 0.873, and clinical NCCT deep learning multilayer perceptron model AUC 0.921. CONCLUSION: The combined clinical imaging deep learning model has a high predictive effect for early HE in sICH patients, which is helpful for clinical individualized assessment of the risk of early HE in sICH patients.


Assuntos
Hemorragia Cerebral , Aprendizado Profundo , Hematoma , Tomografia Computadorizada por Raios X , Humanos , Hemorragia Cerebral/diagnóstico por imagem , Masculino , Hematoma/diagnóstico por imagem , Feminino , Pessoa de Meia-Idade , Idoso , Valor Preditivo dos Testes , Modelos Logísticos , Progressão da Doença , Teorema de Bayes
4.
Sensors (Basel) ; 24(5)2024 Feb 26.
Artigo em Inglês | MEDLINE | ID: mdl-38475050

RESUMO

Latent Low-Rank Representation (LatLRR) has emerged as a prominent approach for fusing visible and infrared images. In this approach, images are decomposed into three fundamental components: the base part, salient part, and sparse part. The aim is to blend the base and salient features to reconstruct images accurately. However, existing methods often focus more on combining the base and salient parts, neglecting the importance of the sparse component, whereas we advocate for the comprehensive inclusion of all three parts generated from LatLRR image decomposition into the image fusion process, a novel proposition introduced in this study. Moreover, the effective integration of Convolutional Neural Network (CNN) technology with LatLRR remains challenging, particularly after the inclusion of sparse parts. This study utilizes fusion strategies involving weighted average, summation, VGG19, and ResNet50 in various combinations to analyze the fusion performance following the introduction of sparse parts. The research findings show a significant enhancement in fusion performance achieved through the inclusion of sparse parts in the fusion process. The suggested fusion strategy involves employing deep learning techniques for fusing both base parts and sparse parts while utilizing a summation strategy for the fusion of salient parts. The findings improve the performance of LatLRR-based methods and offer valuable insights for enhancement, leading to advancements in the field of image fusion.

5.
BMC Bioinformatics ; 24(1): 382, 2023 Oct 10.
Artigo em Inglês | MEDLINE | ID: mdl-37817066

RESUMO

An abnormal growth or fatty mass of cells in the brain is called a tumor. They can be either healthy (normal) or become cancerous, depending on the structure of their cells. This can result in increased pressure within the cranium, potentially causing damage to the brain or even death. As a result, diagnostic procedures such as computed tomography, magnetic resonance imaging, and positron emission tomography, as well as blood and urine tests, are used to identify brain tumors. However, these methods can be labor-intensive and sometimes yield inaccurate results. Instead of these time-consuming methods, deep learning models are employed because they are less time-consuming, require less expensive equipment, produce more accurate results, and are easy to set up. In this study, we propose a method based on transfer learning, utilizing the pre-trained VGG-19 model. This approach has been enhanced by applying a customized convolutional neural network framework and combining it with pre-processing methods, including normalization and data augmentation. For training and testing, our proposed model used 80% and 20% of the images from the dataset, respectively. Our proposed method achieved remarkable success, with an accuracy rate of 99.43%, a sensitivity of 98.73%, and a specificity of 97.21%. The dataset, sourced from Kaggle for training purposes, consists of 407 images, including 257 depicting brain tumors and 150 without tumors. These models could be utilized to develop clinically useful solutions for identifying brain tumors in CT images based on these outcomes.


Assuntos
Neoplasias Encefálicas , Redes Neurais de Computação , Humanos , Neoplasias Encefálicas/diagnóstico por imagem , Tomografia Computadorizada por Raios X , Imageamento por Ressonância Magnética , Encéfalo
6.
Sensors (Basel) ; 23(22)2023 Nov 10.
Artigo em Inglês | MEDLINE | ID: mdl-38005478

RESUMO

In the field of computer vision, hand pose estimation (HPE) has attracted significant attention from researchers, especially in the fields of human-computer interaction (HCI) and virtual reality (VR). Despite advancements in 2D HPE, challenges persist due to hand dynamics and occlusions. Accurate extraction of hand features, such as edges, textures, and unique patterns, is crucial for enhancing HPE. To address these challenges, we propose SDFPoseGraphNet, a novel framework that combines the strengths of the VGG-19 architecture with spatial attention (SA), enabling a more refined extraction of deep feature maps from hand images. By incorporating the Pose Graph Model (PGM), the network adaptively processes these feature maps to provide tailored pose estimations. First Inference Module (FIM) potentials, alongside adaptively learned parameters, contribute to the PGM's final pose estimation. The SDFPoseGraphNet, with its end-to-end trainable design, optimizes across all components, ensuring enhanced precision in hand pose estimation. Our proposed model outperforms existing state-of-the-art methods, achieving an average precision of 7.49% against the Convolution Pose Machine (CPM) and 3.84% in comparison to the Adaptive Graphical Model Network (AGMN).

7.
Sensors (Basel) ; 23(5)2023 Feb 27.
Artigo em Inglês | MEDLINE | ID: mdl-36904845

RESUMO

As a fundamental but difficult topic in computer vision, 3D object segmentation has various applications in medical image analysis, autonomous vehicles, robotics, virtual reality, lithium battery image analysis, etc. In the past, 3D segmentation was performed using hand-made features and design techniques, but these techniques could not generalize to vast amounts of data or reach acceptable accuracy. Deep learning techniques have lately emerged as the preferred method for 3D segmentation jobs as a result of their extraordinary performance in 2D computer vision. Our proposed method used a CNN-based architecture called 3D UNET, which is inspired by the famous 2D UNET that has been used to segment volumetric image data. To see the internal changes of composite materials, for instance, in a lithium battery image, it is necessary to see the flow of different materials and follow the directions analyzing the inside properties. In this paper, a combination of 3D UNET and VGG19 has been used to conduct a multiclass segmentation of publicly available sandstone datasets to analyze their microstructures using image data based on four different objects in the samples of volumetric data. In our image sample, there are a total of 448 2D images, which are then aggregated as one 3D volume to examine the 3D volumetric data. The solution involves the segmentation of each object in the volume data and further analysis of each object to find its average size, area percentage, total area, etc. The open-source image processing package IMAGEJ is used for further analysis of individual particles. In this study, it was demonstrated that convolutional neural networks can be trained to recognize sandstone microstructure traits with an accuracy of 96.78% and an IOU of 91.12%. According to our knowledge, many prior works have applied 3D UNET for segmentation, but very few papers extend it further to show the details of particles in the sample. The proposed solution offers a computational insight for real-time implementation and is discovered to be superior to the current state-of-the-art methods. The result has importance for the creation of an approximately similar model for the microstructural analysis of volumetric data.

8.
Expert Syst Appl ; 211: 118576, 2023 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-36062267

RESUMO

In the last few decades, several epidemic diseases have been introduced. In some cases, doctors and medical physicians are facing difficulties in identifying these diseases correctly. A machine can perform some of these identification tasks more accurately than a human if it is trained correctly. With time, the number of medical data is increasing. A machine can analyze this medical data and extract knowledge from this data, which can help doctors and medical physicians. This study proposed a lightweight convolutional neural network (CNN) named ChestX-ray6 that automatically detects pneumonia, COVID19, cardiomegaly, lung opacity, and pleural from digital chest x-ray images. Here multiple databases have been combined, containing 9,514 chest x-ray images of normal and other five diseases. The lightweight ChestX-ray6 model achieved an accuracy of 80% for the detection of six diseases. The ChestX-ray6 model has been saved and used for binary classification of normal and pneumonia patients to reveal the model's generalization power. The pre-trained ChestX-ray6 model has achieved an accuracy and recall of 97.94% and 98% for binary classification, which outweighs the state-of-the-art (SOTA) models.

9.
Adv Eng Softw ; 175: 103317, 2023 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-36311489

RESUMO

The Coronavirus (COVID-19) has become a critical and extreme epidemic because of its international dissemination. COVID-19 is the world's most serious health, economic, and survival danger. This disease affects not only a single country but the entire planet due to this infectious disease. Illnesses of Covid-19 spread at a much faster rate than usual influenza cases. Because of its high transmissibility and early diagnosis, it isn't easy to manage COVID-19. The popularly used RT-PCR method for COVID-19 disease diagnosis may provide false negatives. COVID-19 can be detected non-invasively using medical imaging procedures such as chest CT and chest x-ray. Deep learning is the most effective machine learning approach for examining a considerable quantity of chest computed tomography (CT) pictures that can significantly affect Covid-19 screening. Convolutional neural network (CNN) is one of the most popular deep learning techniques right now, and its gaining traction due to its potential to transform several spheres of human life. This research aims to develop conceptual transfer learning enhanced CNN framework models for detecting COVID-19 with CT scan images. Though with minimal datasets, these techniques were demonstrated to be effective in detecting the presence of COVID-19. This proposed research looks into several deep transfer learning-based CNN approaches for detecting the presence of COVID-19 in chest CT images.VGG16, VGG19, Densenet121, InceptionV3, Xception, and Resnet50 are the foundation models used in this work. Each model's performance was evaluated using a confusion matrix and various performance measures such as accuracy, recall, precision, f1-score, loss, and ROC. The VGG16 model performed much better than the other models in this study (98.00 % accuracy). Promising outcomes from experiments have revealed the merits of the proposed model for detecting and monitoring COVID-19 patients. This could help practitioners and academics create a tool to help minimal health professionals decide on the best course of therapy.

10.
Comput Electr Eng ; 108: 108711, 2023 May.
Artigo em Inglês | MEDLINE | ID: mdl-37065503

RESUMO

A novel coronavirus (COVID-19), belonging to a family of severe acute respiratory syndrome coronavirus 2 (SARs-CoV-2), was identified in Wuhan city, Hubei, China, in November 2019. The disease had already infected more than 681.529665 million people as of March 13, 2023. Hence, early detection and diagnosis of COVID-19 are essential. For this purpose, radiologists use medical images such as X-ray and computed tomography (CT) images for the diagnosis of COVID-19. It is very difficult for researchers to help radiologists to do automatic diagnoses by using traditional image processing methods. Therefore, a novel artificial intelligence (AI)-based deep learning model to detect COVID-19 from chest X-ray images is proposed. The proposed work uses a wavelet and stacked deep learning architecture (ResNet50, VGG19, Xception, and DarkNet19) named WavStaCovNet-19 to detect COVID-19 from chest X-ray images automatically. The proposed work has been tested on two publicly available datasets and achieved an accuracy of 94.24% and 96.10% on 4 classes and 3 classes, respectively. From the experimental results, we believe that the proposed work can surely be useful in the healthcare domain to detect COVID-19 with less time and cost, and with higher accuracy.

11.
Sensors (Basel) ; 22(17)2022 Sep 02.
Artigo em Inglês | MEDLINE | ID: mdl-36081117

RESUMO

Finding a template in a search image is an important task underlying many computer vision applications. This is typically solved by calculating a similarity map using features extracted from the separate images. Recent approaches perform template matching in a deep feature space, produced by a convolutional neural network (CNN), which is found to provide more tolerance to changes in appearance. Inspired by these findings, in this article we investigate whether enhancing the CNN's encoding of shape information can produce more distinguishable features that improve the performance of template matching. By comparing features from the same CNN trained using different shape-texture training methods, we determined a feature space which improves the performance of most template matching algorithms. When combining the proposed method with the Divisive Input Modulation (DIM) template matching algorithm, its performance is greatly improved, and the resulting method produces state-of-the-art results on a standard benchmark. To confirm these results, we create a new benchmark and show that the proposed method outperforms existing techniques on this new dataset.


Assuntos
Processamento de Imagem Assistida por Computador , Redes Neurais de Computação , Algoritmos , Processamento de Imagem Assistida por Computador/métodos
12.
Sensors (Basel) ; 23(1)2022 Dec 27.
Artigo em Inglês | MEDLINE | ID: mdl-36616876

RESUMO

Brain abnormality causes severe human problems, and thorough screening is necessary to identify the disease. In clinics, bio-image-supported brain abnormality screening is employed mainly because of its investigative accuracy compared with bio-signal (EEG)-based practice. This research aims to develop a reliable disease screening framework for the automatic identification of schizophrenia (SCZ) conditions from brain MRI slices. This scheme consists following phases: (i) MRI slices collection and pre-processing, (ii) implementation of VGG16 to extract deep features (DF), (iii) collection of handcrafted features (HF), (iv) mayfly algorithm-supported optimal feature selection, (v) serial feature concatenation, and (vi) binary classifier execution and validation. The performance of the proposed scheme was independently tested with DF, HF, and concatenated features (DF+HF), and the achieved outcome of this study verifies that the schizophrenia screening accuracy with DF+HF is superior compared with other methods. During this work, 40 patients' brain MRI images (20 controlled and 20 SCZ class) were considered for the investigation, and the following accuracies were achieved: DF provided >91%, HF obtained >85%, and DF+HF achieved >95%. Therefore, this framework is clinically significant, and in the future, it can be used to inspect actual patients' brain MRI slices.


Assuntos
Encefalopatias , Ephemeroptera , Esquizofrenia , Animais , Humanos , Esquizofrenia/diagnóstico por imagem , Imageamento por Ressonância Magnética/métodos , Algoritmos , Encéfalo/diagnóstico por imagem
13.
Sensors (Basel) ; 22(20)2022 Oct 13.
Artigo em Inglês | MEDLINE | ID: mdl-36298129

RESUMO

A color transfer algorithm between images based on two-stage convolutional neural network (CNN) is proposed. The first stage network is based on VGG19 architecture as the backbone. The reference image-based color transfer (RICT) model was used to extract the features of the reference image and the target image, so as to realize the color transfer between them. The second stage is based on progressive convolutional neural network (PCNN) as its backbone. The palette-based emotional color enhancement (PECE) model is adopted to enhance the emotional coloring of the resulting image by comparing the palette, emotional value and the proportion of each color of the reference image. Through five sets of experiments, it is proved that the visual effect processed by our model is obviously better than several main colorization methods in both subjective evaluation and objective data. It can be applied to various complex scenes, and in the near future, it can also be better applied to the fields of digital restoration of old image archives, medical image coloring, art restoration, remote sensing image enhancement, infrared image enhancement and other fields.


Assuntos
Processamento de Imagem Assistida por Computador , Redes Neurais de Computação , Processamento de Imagem Assistida por Computador/métodos , Algoritmos , Aumento da Imagem
14.
Sensors (Basel) ; 22(23)2022 Nov 24.
Artigo em Inglês | MEDLINE | ID: mdl-36501822

RESUMO

The emergence of advanced machine learning or deep learning techniques such as autoencoders and generative adversarial networks, can generate images known as deepfakes, which astonishingly resemble the realistic images. These deepfake images are hard to distinguish from the real images and are being used unethically against famous personalities such as politicians, celebrities, and social workers. Hence, we propose a method to detect these deepfake images using a light weighted convolutional neural network (CNN). Our research is conducted with Deep Fake Detection Challenge (DFDC) full and sample datasets, where we compare the performance of our proposed model with various state-of-the-art pretrained models such as VGG-19, Xception and Inception-ResNet-v2. Furthermore, we perform the experiments with various resolutions maintaining 1:1 and 9:16 aspect ratios, which have not been explored for DFDC datasets by any other groups to date. Thus, the proposed model can flexibly accommodate various resolutions and aspect ratios, without being constrained to a specific resolution or aspect ratio for any type of image classification problem. While most of the reported research is limited to sample or preview DFDC datasets only, we have also attempted the testing on full DFDC datasets and presented the results. Contemplating the fact that the detailed results and resource analysis for various scenarios are provided in this research, the proposed deepfake detection method is anticipated to pave new avenues for deepfake detection research, that engages with DFDC datasets.


Assuntos
Aprendizado de Máquina , Redes Neurais de Computação , Humanos
15.
J Xray Sci Technol ; 30(2): 275-291, 2022.
Artigo em Inglês | MEDLINE | ID: mdl-35001904

RESUMO

Diabetic retinopathy is an eye deficiency that affects retina as a result of the patient having diabetes mellitus caused by high sugar levels, which may eventually lead to macular edema. The objective of this study is to design and compare several deep learning models that detect severity of diabetic retinopathy, determine risk of leading to macular edema, and segment different types of disease patterns using retina images. Indian Diabetic Retinopathy Image Dataset (IDRiD) dataset was used for disease grading and segmentation. Since images of the dataset have different brightness and contrast, we employed three techniques for generating processed images from the original images, which include brightness, color and, contrast (BCC) enhancing, color jitters (CJ), and contrast limited adaptive histogram equalization (CLAHE). After image preporcessing, we used pre-trained ResNet50, VGG16, and VGG19 models on these different preprocessed images both for determining the severity of the retinopathy and also the chances of macular edema. UNet was also applied to segment different types of diseases. To train and test these models, image dataset was divided into training, testing, and validation data at 70%, 20%, and 10% ratios, respectively. During model training, data augmentation method was also applied to increase the number of training images. Study results show that for detecting the severity of retinopathy and macular edema, ResNet50 showed the best accuracy using BCC and original images with an accuracy of 60.2% and 82.5%, respectively, on validation dataset. In segmenting different types of diseases, UNet yielded the highest testing accuracy of 65.22% and 91.09% for microaneurysms and hard exudates using BCC images, 84.83% for optic disc using CJ images, 59.35% and 89.69% for hemorrhages and soft exudates using CLAHE images, respectively. Thus, image preprocessing can play an important role to improve efficacy and performance of deep learning models.


Assuntos
Diabetes Mellitus , Retinopatia Diabética , Edema Macular , Retinopatia Diabética/diagnóstico por imagem , Fundo de Olho , Humanos , Edema Macular/diagnóstico por imagem , Redes Neurais de Computação , Retina/diagnóstico por imagem
16.
Knowl Based Syst ; 258: 110040, 2022 Dec 22.
Artigo em Inglês | MEDLINE | ID: mdl-36284666

RESUMO

During the past two years, a highly infectious virus known as COVID-19 has been damaging and harming the health of people all over the world. Simultaneously, the number of patients is rising in various countries, with many new cases appearing daily, posing a significant challenge to hospital medical staff. It is necessary to improve the efficiency of virus detection. To this end, we combine modern technology and visual assistance to detect COVID-19. Based on the above facts, for accurate and rapid identification of infected persons, the BND-VGG-19 method was proposed. This method is based on VGG-19 and further incorporates batch normalization and dropout layers between the layers to improve network accuracy. Then, the COVID-19 dataset including viral pneumonia, COVID-19, and normal X-ray images, are used to diagnose lung abnormalities and test the performance of the proposed algorithm. The experimental results show the superiority of BND-VGG-19 with a 95.48% accuracy rate compared with existing COVID-19 diagnostic methods.

17.
Sensors (Basel) ; 21(15)2021 Jul 27.
Artigo em Inglês | MEDLINE | ID: mdl-34372306

RESUMO

Accurate identification of siblings through face recognition is a challenging task. This is predominantly because of the high degree of similarities among the faces of siblings. In this study, we investigate the use of state-of-the-art deep learning face recognition models to evaluate their capacity for discrimination between sibling faces using various similarity indices. The specific models examined for this purpose are FaceNet, VGGFace, VGG16, and VGG19. For each pair of images provided, the embeddings have been calculated using the chosen deep learning model. Five standard similarity measures, namely, cosine similarity, Euclidean distance, structured similarity, Manhattan distance, and Minkowski distance, are used to classify images looking for their identity on the threshold defined for each of the similarity measures. The accuracy, precision, and misclassification rate of each model are calculated using standard confusion matrices. Four different experimental datasets for full-frontal-face, eyes, nose, and forehead of sibling pairs are constructed using publicly available HQf subset of the SiblingDB database. The experimental results show that the accuracy of the chosen deep learning models to distinguish siblings based on the full-frontal-face and cropped face areas vary based on the face area compared. It is observed that VGGFace is best while comparing the full-frontal-face and eyes-the accuracy of classification being with more than 95% in this case. However, its accuracy degrades significantly when the noses are compared, while FaceNet provides the best result for classification based on the nose. Similarly, VGG16 and VGG19 are not the best models for classification using the eyes, but these models provide favorable results when foreheads are compared.


Assuntos
Aprendizado Profundo , Reconhecimento Facial , Bases de Dados Factuais , Humanos , Redes Neurais de Computação , Irmãos
18.
Sensors (Basel) ; 21(16)2021 Aug 20.
Artigo em Inglês | MEDLINE | ID: mdl-34451072

RESUMO

Colorectal cancer has become the third most commonly diagnosed form of cancer, and has the second highest fatality rate of cancers worldwide. Currently, optical colonoscopy is the preferred tool of choice for the diagnosis of polyps and to avert colorectal cancer. Colon screening is time-consuming and highly operator dependent. In view of this, a computer-aided diagnosis (CAD) method needs to be developed for the automatic segmentation of polyps in colonoscopy images. This paper proposes a modified SegNet Visual Geometry Group-19 (VGG-19), a form of convolutional neural network, as a CAD method for polyp segmentation. The modifications include skip connections, 5 × 5 convolutional filters, and the concatenation of four dilated convolutions applied in parallel form. The CVC-ClinicDB, CVC-ColonDB, and ETIS-LaribPolypDB databases were used to evaluate the model, and it was found that our proposed polyp segmentation model achieved an accuracy, sensitivity, specificity, precision, mean intersection over union, and dice coefficient of 96.06%, 94.55%, 97.56%, 97.48%, 92.3%, and 95.99%, respectively. These results indicate that our model performs as well as or better than previous schemes in the literature. We believe that this study will offer benefits in terms of the future development of CAD tools for polyp segmentation for colorectal cancer diagnosis and management. In the future, we intend to embed our proposed network into a medical capsule robot for practical usage and try it in a hospital setting with clinicians.


Assuntos
Colonoscopia , Redes Neurais de Computação , Bases de Dados Factuais , Diagnóstico por Computador , Processamento de Imagem Assistida por Computador , Projetos de Pesquisa
19.
Sensors (Basel) ; 21(24)2021 Dec 08.
Artigo em Inglês | MEDLINE | ID: mdl-34960274

RESUMO

The recent development in the area of IoT technologies is likely to be implemented extensively in the next decade. There is a great increase in the crime rate, and the handling officers are responsible for dealing with a broad range of cyber and Internet issues during investigation. IoT technologies are helpful in the identification of suspects, and few technologies are available that use IoT and deep learning together for face sketch synthesis. Convolutional neural networks (CNNs) and other constructs of deep learning have become major tools in recent approaches. A new-found architecture of the neural network is anticipated in this work. It is called Spiral-Net, which is a modified version of U-Net fto perform face sketch synthesis (the phase is known as the compiler network C here). Spiral-Net performs in combination with a pre-trained Vgg-19 network called the feature extractor F. It first identifies the top n matches from viewed sketches to a given photo. F is again used to formulate a feature map based on the cosine distance of a candidate sketch formed by C from the top n matches. A customized CNN configuration (called the discriminator D) then computes loss functions based on differences between the candidate sketch and the feature. Values of these loss functions alternately update C and F. The ensemble of these nets is trained and tested on selected datasets, including CUFS, CUFSF, and a part of the IIT photo-sketch dataset. Results of this modified U-Net are acquired by the legacy NLDA (1998) scheme of face recognition and its newer version, OpenBR (2013), which demonstrate an improvement of 5% compared with the current state of the art in its relevant domain.


Assuntos
Inteligência Artificial , Aprendizado Profundo , Algoritmos , Face , Redes Neurais de Computação
20.
Pattern Recognit Lett ; 152: 122-128, 2021 Dec.
Artigo em Inglês | MEDLINE | ID: mdl-34566222

RESUMO

COVID-19 is an infectious and contagious virus. As of this writing, more than 160 million people have been infected since its emergence, including more than 125,000 in Algeria. In this work, We first collected a dataset of 4986 COVID and non-COVID images confirmed by RT-PCR tests at Tlemcen hospital in Algeria. Then we performed a transfer learning on deep learning models that got the best results on the ImageNet dataset, such as DenseNet121, DenseNet201, VGG16, VGG19, Inception Resnet-V2, and Xception, in order to conduct a comparative study. Therefore, We have proposed an explainable model based on the DenseNet201 architecture and the GradCam explanation algorithm to detect COVID-19 in chest CT images and explain the output decision. Experiments have shown promising results and proven that the introduced model can be beneficial for diagnosing and following up patients with COVID-19.

SELEÇÃO DE REFERÊNCIAS
Detalhe da pesquisa