Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 9 de 9
Filtrar
1.
J Digit Imaging ; 36(2): 627-646, 2023 04.
Artículo en Inglés | MEDLINE | ID: mdl-36515746

RESUMEN

Breast ultrasound (BUS) imaging has become one of the key imaging modalities for medical image diagnosis and prognosis. However, the manual process of lesion delineation from ultrasound images can incur various challenges concerning variable shape, size, intensity, curvature, or other medical priors of the lesion in the image. Therefore, computer-aided diagnostic (CADx) techniques incorporating deep learning-based neural networks are automatically used to segment the lesion from BUS images. This paper proposes an encoder-decoder-based architecture to recognize and accurately segment the lesion from two-dimensional BUS images. The architecture is utilized with the residual connection in both encoder and decoder paths; bi-directional ConvLSTM (BConvLSTM) units in the decoder extract the minute and detailed region of interest (ROI) information. BConvLSTM units and residual blocks help the network weigh ROI information more than the similar background region. Two public BUS image datasets, one with 163 images and the other with 42 images, are used. The proposed model is trained with the augmented images (ten forms) of dataset one (with 163 images), and test results are produced on the second dataset and the testing set of the first dataset-the segmentation performance yielding comparable results with the state-of-the-art segmentation methodologies. Similarly, the visual results show that the proposed approach for BUS image segmentation can accurately identify lesion contours and can potentially be applied for similar and larger datasets.


Asunto(s)
Neoplasias de la Mama , Procesamiento de Imagen Asistido por Computador , Humanos , Femenino , Procesamiento de Imagen Asistido por Computador/métodos , Mama/diagnóstico por imagen , Redes Neurales de la Computación , Ultrasonografía , Neoplasias de la Mama/diagnóstico por imagen
2.
Med Biol Eng Comput ; 62(7): 2037-2058, 2024 Jul.
Artículo en Inglés | MEDLINE | ID: mdl-38436836

RESUMEN

This paper introduces a novel approach to enhance content-based image retrieval, validated on two benchmark datasets: ISIC-2017 and ISIC-2018. These datasets comprise skin lesion images that are crucial for innovations in skin cancer diagnosis and treatment. We advocate the use of pre-trained Vision Transformer (ViT), a relatively uncharted concept in the realm of image retrieval, particularly in medical scenarios. In contrast to the traditionally employed Convolutional Neural Networks (CNNs), our findings suggest that ViT offers a more comprehensive understanding of the image context, essential in medical imaging. We further incorporate a weighted multi-loss function, delving into various losses such as triplet loss, distillation loss, contrastive loss, and cross-entropy loss. Our exploration investigates the most resilient combination of these losses to create a robust multi-loss function, thus enhancing the robustness of the learned feature space and ameliorating the precision and recall in the retrieval process. Instead of using all the loss functions, the proposed multi-loss function utilizes the combination of only cross-entropy loss, triplet loss, and distillation loss and gains improvement of 6.52% and 3.45% for mean average precision over ISIC-2017 and ISIC-2018. Another innovation in our methodology is a two-branch network strategy, which concurrently boosts image retrieval and classification. Through our experiments, we underscore the effectiveness and the pitfalls of diverse loss configurations in image retrieval. Furthermore, our approach underlines the advantages of retrieval-based classification through majority voting rather than relying solely on the classification head, leading to enhanced prediction for melanoma - the most lethal type of skin cancer. Our results surpass existing state-of-the-art techniques on the ISIC-2017 and ISIC-2018 datasets by improving mean average precision by 1.01% and 4.36% respectively, emphasizing the efficacy and promise of Vision Transformers paired with our tailor-made weighted loss function, especially in medical contexts. The proposed approach's effectiveness is substantiated through thorough ablation studies and an array of quantitative and qualitative outcomes. To promote reproducibility and support forthcoming research, our source code will be accessible on GitHub.


Asunto(s)
Redes Neurales de la Computación , Humanos , Neoplasias Cutáneas/diagnóstico por imagen , Neoplasias Cutáneas/patología , Algoritmos , Procesamiento de Imagen Asistido por Computador/métodos , Bases de Datos Factuales , Interpretación de Imagen Asistida por Computador/métodos
3.
Neural Netw ; 150: 392-407, 2022 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-35358887

RESUMEN

In this paper, a novel dual-channel system for multi-class text emotion recognition has been proposed, and a novel technique to explain its training & predictions has been developed. The architecture of the proposed system contains the embedding module, dual-channel module, emotion classification module, and explainability module. The embedding module extracts the textual features from the input sentences in the form of embedding vectors using the pre-trained Bidirectional Encoder Representations from Transformers (BERT) model. Then the embedding vectors are fed as the inputs to the dual-channel network containing two network channels made up of convolutional neural network (CNN) and bidirectional long short term memory (BiLSTM) network. The intuition behind using CNN and BiLSTM in both the channels was to harness the goodness of the convolutional layer for feature extraction and the BiLSTM layer to extract text's order and sequence-related information. The outputs of both channels are in the form of embedding vectors which are concatenated and fed to the emotion classification module. The proposed system's architecture has been determined by thorough ablation studies, and a framework has been developed to discuss its computational cost. The emotion classification module learns and projects the emotion embeddings on a hyperplane in the form of clusters. The proposed explainability technique explains the training and predictions of the proposed system by analyzing the inter & intra-cluster distances and the intersection of these clusters. The proposed approach's consistent accuracy, precision, recall, and F1 score results for ISEAR, Aman, AffectiveText, and EmotionLines datasets, ensure its applicability to diverse texts.


Asunto(s)
Lenguaje , Redes Neurales de la Computación , Recolección de Datos , Emociones
4.
SN Comput Sci ; 3(1): 41, 2022.
Artículo en Inglés | MEDLINE | ID: mdl-34746807

RESUMEN

The sudden advent of COVID-19 pandemic left educational institutions in a difficult situation for the semester evaluation of students; especially where the online participation was difficult for the students. Such a situation may also happen during a similar disaster in the future. Through this work, we want to study the question: can the deep learning methods be leveraged to predict student grades based on the available performance of students. To this end, this paper presents an in-depth analysis of deep learning and machine learning approaches for the formulation of an automated students' performance estimation system that works on partially available students' academic records. Our main contributions are: (a) a large dataset with 15 courses (shared publicly for academic research); (b) statistical analysis and ablations on the estimation problem for this dataset; (c) predictive analysis through deep learning approaches and comparison with other arts and machine learning algorithms. Unlike previous approaches that rely on feature engineering or logical function deduction, our approach is fully data-driven and thus highly generic with better performance across different prediction tasks. The main takeaways from this study are: (a) for better prediction rates, it is desirable to have multiple low weightage tests than few very high weightage exams; (b) the latent space models are better estimators than sequential models; (c) deep learning models have the potential to very accurately estimate the student performance and their accuracy only improves as the training data are increased.

5.
Multimed Tools Appl ; 81(19): 27631-27655, 2022.
Artículo en Inglés | MEDLINE | ID: mdl-35368858

RESUMEN

COVID-19 is a viral disease that in the form of a pandemic has spread in the entire world, causing a severe impact on people's well being. In fighting against this deadly disease, a pivotal step can prove to be an effective screening and diagnosing step to treat infected patients. This can be made possible through the use of chest X-ray images. Early detection using the chest X-ray images can prove to be a key solution in fighting COVID-19. Many computer-aided diagnostic (CAD) techniques have sprung up to aid radiologists and provide them a secondary suggestion for the same. In this study, we have proposed the notion of Pearson Correlation Coefficient (PCC) along with variance thresholding to optimally reduce the feature space of extracted features from the conventional deep learning architectures, ResNet152 and GoogLeNet. Further, these features are classified using machine learning (ML) predictive classifiers for multi-class classification among COVID-19, Pneumonia and Normal. The proposed model is validated and tested on publicly available COVID-19 and Pneumonia and Normal dataset containing an extensive set of 768 images of COVID-19 with 5216 training images of Pneumonia and Normal patients. Experimental results reveal that the proposed model outperforms other previous related works. While the achieved results are encouraging, further analysis on the COVID-19 images can prove to be more reliable for effective classification.

6.
Phys Eng Sci Med ; 44(4): 1257-1271, 2021 Dec.
Artículo en Inglés | MEDLINE | ID: mdl-34609703

RESUMEN

According to the World Health Organization (WHO), novel coronavirus (COVID-19) is an infectious disease and has a significant social and economic impact. The main challenge in fighting against this disease is its scale. Due to the outbreak, medical facilities are under pressure due to case numbers. A quick diagnosis system is required to address these challenges. To this end, a stochastic deep learning model is proposed. The main idea is to constrain the deep-representations over a Gaussian prior to reinforce the discriminability in feature space. The model can work on chest X-ray or CT-scan images. It provides a fast diagnosis of COVID-19 and can scale seamlessly. The work presents a comprehensive evaluation of previously proposed approaches for X-ray based disease diagnosis. The approach works by learning a latent space over X-ray image distribution from the ensemble of state-of-the-art convolutional-nets, and then linearly regressing the predictions from an ensemble of classifiers which take the latent vector as input. We experimented with publicly available datasets having three classes: COVID-19, normal and pneumonia yielding an overall accuracy and AUC of 0.91 and 0.97, respectively. Moreover, for robust evaluation, experiments were performed on a large chest X-ray dataset to classify among Atelectasis, Effusion, Infiltration, Nodule, and Pneumonia classes. The results demonstrate that the proposed model has better understanding of the X-ray images which make the network more generic to be later used with other domains of medical image analysis.


Asunto(s)
COVID-19 , Aprendizaje Profundo , Algoritmos , Humanos , Redes Neurales de la Computación , SARS-CoV-2 , Rayos X
7.
Med Biol Eng Comput ; 58(6): 1199-1211, 2020 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-32200453

RESUMEN

Breast cancer has the second highest frequency of death rate among women worldwide. Early-stage prevention becomes complex due to reasons unknown. However, some typical signatures like masses and micro-calcifications upon investigating mammograms can help diagnose women better. Manual diagnosis is a hard task the radiologists carry out frequently. For their assistance, many computer-aided diagnosis (CADx) approaches have been developed. To improve upon the state of the art, we proposed a deep ensemble transfer learning and neural network classifier for automatic feature extraction and classification. In computer-assisted mammography, deep learning-based architectures are generally not trained on mammogram images directly. Instead, the images are pre-processed beforehand, and then they are adopted to be given as input to the ensemble model proposed. The robust features extracted from the ensemble model are optimized into a feature vector which are further classified using the neural network (nntraintool). The network was trained and tested to separate out benign and malignant tumors, thus achieving an accuracy of 0.88 with an area under curve (AUC) of 0.88. The attained results show that the proposed methodology is a promising and robust CADx system for breast cancer classification. Graphical Abstract Flow diagram of the proposed approach. Figure depicts the deep ensemble extracting the robust features with the final classification using neural networks.


Asunto(s)
Neoplasias de la Mama/diagnóstico por imagen , Diagnóstico por Computador/métodos , Procesamiento de Imagen Asistido por Computador/métodos , Mamografía/métodos , Algoritmos , Área Bajo la Curva , Enfermedades de la Mama/diagnóstico por imagen , Calcinosis/diagnóstico por imagen , Aprendizaje Profundo , Femenino , Humanos , Mamografía/clasificación , Redes Neurales de la Computación , Máquina de Vectores de Soporte
8.
Neural Netw ; 92: 77-88, 2017 Aug.
Artículo en Inglés | MEDLINE | ID: mdl-28254237

RESUMEN

This paper presents a novel approach to predict rating of video-advertisements based on a multimodal framework combining physiological analysis of the user and global sentiment-rating available on the internet. We have fused Electroencephalogram (EEG) waves of user and corresponding global textual comments of the video to understand the user's preference more precisely. In our framework, the users were asked to watch the video-advertisement and simultaneously EEG signals were recorded. Valence scores were obtained using self-report for each video. A higher valence corresponds to intrinsic attractiveness of the user. Furthermore, the multimedia data that comprised of the comments posted by global viewers, were retrieved and processed using Natural Language Processing (NLP) technique for sentiment analysis. Textual contents from review comments were analyzed to obtain a score to understand sentiment nature of the video. A regression technique based on Random forest was used to predict the rating of an advertisement using EEG data. Finally, EEG based rating is combined with NLP-based sentiment score to improve the overall prediction. The study was carried out using 15 video clips of advertisements available online. Twenty five participants were involved in our study to analyze our proposed system. The results are encouraging and these suggest that the proposed multimodal approach can achieve lower RMSE in rating prediction as compared to the prediction using only EEG data.


Asunto(s)
Publicidad , Encéfalo/fisiología , Comportamiento del Consumidor , Electroencefalografía/métodos , Modelos Neurológicos , Humanos , Internet , Procesamiento de Lenguaje Natural
9.
ISA Trans ; 53(2): 547-59, 2014 Mar.
Artículo en Inglés | MEDLINE | ID: mdl-24388356

RESUMEN

This paper proposes a novel phase based approach for computing disparity as the optical flow from the given pair of consecutive images. A new dual tree fractional quaternion wavelet transform (FrQWT) is proposed by defining the 2D Fourier spectrum upto a single quadrant. In the proposed FrQWT, each quaternion wavelet consists of a real part (a real DWT wavelet) and three imaginary parts that are organized according to the quaternion algebra. First two FrQWT phases encode the shifts of image features in the absolute horizontal and vertical coordinate system, while the third phase has the texture information. The FrQWT allowed a multi-scale framework for calculating and adjusting local disparities and executing phase unwrapping from coarse to fine scales with linear computational efficiency.


Asunto(s)
Procesamiento de Imagen Asistido por Computador/métodos , Visión Ocular , Algoritmos , Simulación por Computador , Análisis de Fourier , Interpretación de Imagen Asistida por Computador/métodos , Distribución Normal , Análisis de Ondículas
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA