Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 39
Filtrar
1.
Neural Netw ; 180: 106572, 2024 Aug 08.
Artigo em Inglês | MEDLINE | ID: mdl-39173200

RESUMO

Person Re-identification (Re-ID) aims to match person images across non-overlapping cameras. The existing approaches formulate this task as fine-grained representation learning with deep neural networks, which involves extracting image features using a deep convolutional network, followed by mapping the features into a discriminative space through another smaller network, in order to make full use of all possible cues. However, recent Re-ID methods that strive to capture every cue and make the space more discriminative have resulted in longer features, ranging from 1024 to 14336, leading to higher time (distance computation) and space (feature storage) complexities. There are two potential solutions: reduction-after-training methods (such as Principal Component Analysis and Linear Discriminant Analysis) and reduction-during-training methods (such as 1 × 1 Convolution). The former utilizes a statistical approach aiming for a global optimum but lacking end-to-end optimization of large data and deep neural networks. The latter lacks theoretical guarantees and may be vulnerable to training noise such as dataset noise or initialization seed. To address these limitations, we propose a method called Euclidean-Distance-Preserving Feature Reduction (EDPFR) that combines the strengths of both reduction-after-training and reduction-during-training methods. EDPFR first formulates the feature reduction process as a matrix decomposition and derives a condition to preserve the Euclidean distance between features, thus ensuring accuracy in theory. Furthermore, the method integrates the matrix decomposition process into a deep neural network to enable end-to-end optimization and batch training, while maintaining the theoretical guarantee. The result of the EDPFR is a reduction of the feature dimensions from fa and fb to fa' and fb', while preserving their Euclidean distance, i.e.L2(fa,fb)=L2(fa',fb'). In addition to its Euclidean-Distance-Preserving capability, EDPFR also features a novel feature-level distillation loss. One of the main challenges in knowledge distillation is dimension mismatch. While previous distillation losses, usually project the mismatched features to matched class-level, spatial-level, or similarity-level spaces, this can result in a loss of information and decrease the flexibility and efficiency of distillation. Our proposed feature-level distillation leverages the benefits of the Euclidean-Distance-Preserving property and performs distillation directly in the feature space, resulting in a more flexible and efficient approach. Extensive on three Re-ID datasets, Market-1501, DukeMTMC-reID and MSMT demonstrate the effectiveness of our proposed Euclidean-Distance-Preserving Feature Reduction.

2.
Artigo em Inglês | MEDLINE | ID: mdl-38980780

RESUMO

Graph neural networks (GNNs), especially dynamic GNNs, have become a research hotspot in spatiotemporal forecasting problems. While many dynamic graph construction methods have been developed, relatively few of them explore the causal relationship between neighbor nodes. Thus, the resulting models lack strong explainability for the causal relationship between the neighbor nodes of the dynamically generated graphs, which can easily lead to a risk in subsequent decisions. Moreover, few of them consider the uncertainty and noise of dynamic graphs based on the time series datasets, which are ubiquitous in real-world graph structure networks. In this article, we propose a novel dynamic diffusion-variational GNN (DVGNN) for spatiotemporal forecasting. For dynamic graph construction, an unsupervised generative model is devised. Two layers of graph convolutional network (GCN) are applied to calculate the posterior distribution of the latent node embeddings in the encoder stage. Then, a diffusion model is used to infer the dynamic link probability and reconstruct causal graphs (CGs) in the decoder stage adaptively. The new loss function is derived theoretically, and the reparameterization trick is adopted in estimating the probability distribution of the dynamic graphs by evidence lower bound (ELBO) during the backpropagation period. After obtaining the generated graphs, dynamic GCN and temporal attention are applied to predict future states. Experiments are conducted on four real-world datasets of different graph structures in different domains. The results demonstrate that the proposed DVGNN model outperforms state-of-the-art approaches and achieves outstanding root mean square error (RMSE) results while exhibiting higher robustness. Also, by F1-score and probability distribution analysis, we demonstrate that DVGNN better reflects the causal relationship and uncertainty of dynamic graphs. The website of the code is https://github.com/gorgen2020/DVGNN.

3.
Neural Netw ; 178: 106458, 2024 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-38901093

RESUMO

The detection of therapeutic peptides is a topic of immense interest in the biomedical field. Conventional biochemical experiment-based detection techniques are tedious and time-consuming. Computational biology has become a useful tool for improving the detection efficiency of therapeutic peptides. Most computational methods do not consider the deviation caused by noise. To improve the generalization performance of therapeutic peptide prediction methods, this work presents a sequence homology score-based deep fuzzy echo-state network with maximizing mixture correntropy (SHS-DFESN-MMC) model. Our method is compared with the existing methods on eight types of therapeutic peptide datasets. The model parameters are determined by 10 fold cross-validation on their training sets and verified by independent test sets. Across the 8 datasets, the average area under the receiver operating characteristic curve (AUC) values of SHS-DFESN-MMC are the highest on both the training (0.926) and independent sets (0.923).


Assuntos
Lógica Fuzzy , Redes Neurais de Computação , Peptídeos , Biologia Computacional/métodos , Humanos , Aprendizado Profundo , Área Sob a Curva , Curva ROC , Algoritmos
4.
Brief Bioinform ; 25(3)2024 Mar 27.
Artigo em Inglês | MEDLINE | ID: mdl-38622357

RESUMO

Pseudouridine is an RNA modification that is widely distributed in both prokaryotes and eukaryotes, and plays a critical role in numerous biological activities. Despite its importance, the precise identification of pseudouridine sites through experimental approaches poses significant challenges, requiring substantial time and resources.Therefore, there is a growing need for computational techniques that can reliably and quickly identify pseudouridine sites from vast amounts of RNA sequencing data. In this study, we propose fuzzy kernel evidence Random Forest (FKeERF) to identify pseudouridine sites. This method is called PseU-FKeERF, which demonstrates high accuracy in identifying pseudouridine sites from RNA sequencing data. The PseU-FKeERF model selected four RNA feature coding schemes with relatively good performance for feature combination, and then input them into the newly proposed FKeERF method for category prediction. FKeERF not only uses fuzzy logic to expand the original feature space, but also combines kernel methods that are easy to interpret in general for category prediction. Both cross-validation tests and independent tests on benchmark datasets have shown that PseU-FKeERF has better predictive performance than several state-of-the-art methods. This new method not only improves the accuracy of pseudouridine site identification, but also provides a certain reference for disease control and related drug development in the future.


Assuntos
Pseudouridina , Algoritmo Florestas Aleatórias , Pseudouridina/genética , RNA/genética , Sequência de Bases
5.
Comput Biol Med ; 172: 108282, 2024 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-38503085

RESUMO

Cardiac ultrasound (US) image segmentation is vital for evaluating clinical indices, but it often demands a large dataset and expert annotations, resulting in high costs for deep learning algorithms. To address this, our study presents a framework utilizing artificial intelligence generation technology to produce multi-class RGB masks for cardiac US image segmentation. The proposed approach directly performs semantic segmentation of the heart's main structures in US images from various scanning modes. Additionally, we introduce a novel learning approach based on conditional generative adversarial networks (CGAN) for cardiac US image segmentation, incorporating a conditional input and paired RGB masks. Experimental results from three cardiac US image datasets with diverse scan modes demonstrate that our approach outperforms several state-of-the-art models, showcasing improvements in five commonly used segmentation metrics, with lower noise sensitivity. Source code is available at https://github.com/energy588/US2mask.


Assuntos
Inteligência Artificial , Ecocardiografia , Algoritmos , Benchmarking , Semântica , Processamento de Imagem Assistida por Computador
6.
IEEE J Biomed Health Inform ; 28(8): 4494-4502, 2024 Aug.
Artigo em Inglês | MEDLINE | ID: mdl-38261491

RESUMO

Cognitive computing endeavors to construct models that emulate brain functions, which can be explored through electroencephalography (EEG). Developing precise and robust EEG classification models is crucial for advancing cognitive computing. Despite the high accuracy of supervised EEG classification models, they are constrained by labor-intensive annotations and poor generalization. Self-supervised models address these issues but encounter difficulties in matching the accuracy of supervised learning. Three challenges persist: 1) capturing temporal dependencies in EEG; 2) adapting loss functions to describe feature similarities in self-supervised models; and 3) addressing the prevalent issue of data imbalance in EEG. This study introduces the DreamCatcher Network (DCNet), a self-supervised EEG classification framework with a two-stage training strategy. The first stage extracts robust representations through contrastive learning, and the second stage transfers the representation encoder to a supervised EEG classification task. DCNet utilizes time-series contrastive learning to autonomously construct representations that comprehensively capture temporal correlations. A novel loss function, SelfDreamCatcherLoss, is proposed to evaluate the similarities between these representations and enhance the performance of DCNet. Additionally, two data augmentation methods are integrated to alleviate class imbalances. Extensive experiments show the superiority of DCNet over the current state-of-the-art models, achieving high accuracy on both the Sleep-EDF and HAR datasets. It holds substantial promise for revolutionizing sleep disorder detection and expediting the development of advanced healthcare systems driven by cognitive computing.


Assuntos
Eletroencefalografia , Processamento de Sinais Assistido por Computador , Aprendizado de Máquina Supervisionado , Humanos , Eletroencefalografia/métodos , Eletroencefalografia/classificação , Algoritmos , Redes Neurais de Computação , Encéfalo/fisiologia , Encéfalo/fisiopatologia
7.
Neural Netw ; 171: 14-24, 2024 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-38091757

RESUMO

Document-level relation extraction faces two often overlooked challenges: long-tail problem and multi-label problem. Previous work focuses mainly on obtaining better contextual representations for entity pairs, hardly address the above challenges. In this paper, we analyze the co-occurrence correlation of relations, and introduce it into the document-level relation extraction task for the first time. We argue that the correlations can not only transfer knowledge between data-rich relations and data-scarce ones to assist in the training of long-tailed relations, but also reflect semantic distance guiding the classifier to identify semantically close relations for multi-label entity pairs. Specifically, we use relation embedding as a medium, and propose two co-occurrence prediction sub-tasks from both coarse- and fine-grained perspectives to capture relation correlations. Finally, the learned correlation-aware embeddings are used to guide the extraction of relational facts. Substantial experiments on two popular datasets (i.e., DocRED and DWIE) are conducted, and our method achieves superior results compared to baselines. Insightful analysis also demonstrates the potential of relation correlations to address the above challenges. The data and code are released at https://github.com/RidongHan/DocRE-Co-Occur.


Assuntos
Semântica
8.
Neural Netw ; 169: 307-324, 2024 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-37922714

RESUMO

Large deep learning models are impressive, but they struggle when real-time data is not available. Few-shot class-incremental learning (FSCIL) poses a significant challenge for deep neural networks to learn new tasks from just a few labeled samples without forgetting the previously learned ones. This setup can easily leads to catastrophic forgetting and overfitting problems, severely affecting model performance. Studying FSCIL helps overcome deep learning model limitations on data volume and acquisition time, while improving practicality and adaptability of machine learning models. This paper provides a comprehensive survey on FSCIL. Unlike previous surveys, we aim to synthesize few-shot learning and incremental learning, focusing on introducing FSCIL from two perspectives, while reviewing over 30 theoretical research studies and more than 20 applied research studies. From the theoretical perspective, we provide a novel categorization approach that divides the field into five subcategories, including traditional machine learning methods, meta learning-based methods, feature and feature space-based methods, replay-based methods, and dynamic network structure-based methods. We also evaluate the performance of recent theoretical research on benchmark datasets of FSCIL. From the application perspective, FSCIL has achieved impressive achievements in various fields of computer vision such as image classification, object detection, and image segmentation, as well as in natural language processing and graph. We summarize the important applications. Finally, we point out potential future research directions, including applications, problem setups, and theory development. Overall, this paper offers a comprehensive analysis of the latest advances in FSCIL from a methodological, performance, and application perspective.


Assuntos
Aprendizado de Máquina , Redes Neurais de Computação , Inquéritos e Questionários , Tempo
9.
Neural Netw ; 169: 623-636, 2024 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-37976593

RESUMO

The accurate prediction of drug-target affinity (DTA) is a crucial step in drug discovery and design. Traditional experiments are very expensive and time-consuming. Recently, deep learning methods have achieved notable performance improvements in DTA prediction. However, one challenge for deep learning-based models is appropriate and accurate representations of drugs and targets, especially the lack of effective exploration of target representations. Another challenge is how to comprehensively capture the interaction information between different instances, which is also important for predicting DTA. In this study, we propose AttentionMGT-DTA, a multi-modal attention-based model for DTA prediction. AttentionMGT-DTA represents drugs and targets by a molecular graph and binding pocket graph, respectively. Two attention mechanisms are adopted to integrate and interact information between different protein modalities and drug-target pairs. The experimental results showed that our proposed model outperformed state-of-the-art baselines on two benchmark datasets. In addition, AttentionMGT-DTA also had high interpretability by modeling the interaction strength between drug atoms and protein residues. Our code is available at https://github.com/JK-Liu7/AttentionMGT-DTA.


Assuntos
Benchmarking , Descoberta de Drogas
10.
Comput Biol Med ; 166: 107549, 2023 Sep 30.
Artigo em Inglês | MEDLINE | ID: mdl-37839222

RESUMO

To address the scarcity and class imbalance of abnormal electrocardiogram (ECG) databases, which are crucial in AI-driven diagnostic tools for potential cardiovascular disease detection, this study proposes a novel quantum conditional generative adversarial algorithm (QCGAN-ECG) for generating abnormal ECG signals. The QCGAN-ECG constructs a quantum generator based on patch method. In this method, each sub-generator generates distinct features of abnormal heartbeats in different segments. This patch-based generative algorithm conserves quantum resources and makes QCGAN-ECG practical for near-term quantum devices. Additionally, QCGAN-ECG introduces quantum registers as control conditions. It encodes information about the types and probability distributions of abnormal heartbeats into quantum registers, rendering the entire generative process controllable. Simulation experiments on Pennylane demonstrated that the QCGAN-ECG could generate completely abnormal heartbeats with an average accuracy of 88.8%. Moreover, the QCGAN-ECG can accurately fit the probability distribution of various abnormal ECG data. In the anti-noise experiments, the QCGAN-ECG showcased outstanding robustness across various levels of quantum noise interference. These results demonstrate the effectiveness and potential applicability of the QCGAN-ECG for generating abnormal ECG signals, which will further promote the development of AI-driven cardiac disease diagnosis systems. The source code is available at github.com/VanSWK/QCGAN_ECG.

11.
Artigo em Inglês | MEDLINE | ID: mdl-37552590

RESUMO

Smart healthcare aims to revolutionize med-ical services by integrating artificial intelligence (AI). The limitations of classical machine learning include privacy concerns that prevent direct data sharing among medical institutions, untimely updates, and long training times. To address these issues, this study proposes a digital twin-assisted quantum federated learning algorithm (DTQFL). By leveraging the 5G mobile network, digital twins (DT) of patients can be created instantly using data from various Internet of Medical Things (IoMT) devices and simultane-ously reduce communication time in federated learning (FL) at the same time. DTQFL generates DT for patients with specific diseases, allowing for synchronous training and updating of the variational quantum neural network (VQNN) without disrupting the VQNN in the real world. This study utilized DTQFL to train its own personalized VQNN for each hospital, considering privacy security and training speed. Simultaneously, the personalized VQNN of each hospital was obtained through further local iterations of the final global parameters. The results indicate that DTQFL can train a good VQNN without collecting local data while achieving accuracy comparable to that of data-centralized algorithms. In addition, after personalized train-ing, the VQNN can achieve higher accuracy than that with-out personalized training.

12.
Artigo em Inglês | MEDLINE | ID: mdl-37610905

RESUMO

Given the overwhelming and rapidly increasing volumes of the published biomedical literature, automatic biomedical text summarization has long been a highly important task. Recently, great advances in the performance of biomedical text summarization have been facilitated by pre-trained language models (PLMs) based on fine-tuning. However, existing summarization methods based on PLMs do not capture domain-specific knowledge. This can result in generated summaries with low coherence, including redundant sentences, or excluding important domain knowledge conveyed in the full-text document. Furthermore, the black-box nature of the transformers means that they lack explainability, i.e. it is not clear to users how and why the summary was generated. The domain-specific knowledge and explainability are crucial for the accuracy and transparency of biomedical text summarization methods. In this article, we aim to address these issues by proposing a novel domain knowledge-enhanced graph topic transformer (DORIS) for explainable biomedical text summarization. The model integrates the graph neural topic model and the domain-specific knowledge from the Unified Medical Language System (UMLS) into the transformer-based PLM, to improve the explainability and accuracy. Experimental results on four biomedical literature datasets show that our model outperforms existing state-of-the-art (SOTA) PLM-based summarization methods on biomedical extractive summarization. Furthermore, our use of graph neural topic modeling means that our model possesses the desirable property of being explainable, i.e. it is straightforward for users to understand how and why the model selects particular sentences for inclusion in the summary. The domain-specific knowledge helps our model to learn more coherent topics, to better explain the performance.

13.
Neural Netw ; 166: 188-203, 2023 Sep.
Artigo em Inglês | MEDLINE | ID: mdl-37499604

RESUMO

Subspace distance is an invaluable tool exploited in a wide range of feature selection methods. The power of subspace distance is that it can identify a representative subspace, including a group of features that can efficiently approximate the space of original features. On the other hand, employing intrinsic statistical information of data can play a significant role in a feature selection process. Nevertheless, most of the existing feature selection methods founded on the subspace distance are limited in properly fulfilling this objective. To pursue this void, we propose a framework that takes a subspace distance into account which is called "Variance-Covariance subspace distance". The approach gains advantages from the correlation of information included in the features of data, thus determines all the feature subsets whose corresponding Variance-Covariance matrix has the minimum norm property. Consequently, a novel, yet efficient unsupervised feature selection framework is introduced based on the Variance-Covariance distance to handle both the dimensionality reduction and subspace learning tasks. The proposed framework has the ability to exclude those features that have the least variance from the original feature set. Moreover, an efficient update algorithm is provided along with its associated convergence analysis to solve the optimization side of the proposed approach. An extensive number of experiments on nine benchmark datasets are also conducted to assess the performance of our method from which the results demonstrate its superiority over a variety of state-of-the-art unsupervised feature selection methods. The source code is available at https://github.com/SaeedKarami/VCSDFS.


Assuntos
Algoritmos , Reconhecimento Automatizado de Padrão , Reconhecimento Automatizado de Padrão/métodos , Aprendizagem , Software , Benchmarking
14.
Artigo em Inglês | MEDLINE | ID: mdl-37399158

RESUMO

Electrocardiogram (ECG) is the main criterion for arrhythmia detection. As a means of identification, ECG leakage seems to be a common occurrence due to the development of the Internet of Medical Things (IoMT). The advent of the quantum era makes it difficult for classical blockchain technology to provide security for ECG data storage. Therefore, from the perspective of safety and practicality, this article proposes a quantum arrhythmia detection system named QADS, which achieves secure storage and sharing of ECG data based on quantum blockchain technology. Furthermore, a quantum neural network is used in QADS to recognize abnormal ECG data, which contributes to further cardiovascular disease diagnosis. Each quantum block stores the hash of the current and previous block to construct a quantum block network. The new quantum blockchain algorithm introduces a controlled quantum walk hash function and a quantum authentication protocol to guarantee legitimacy and security while creating new blocks. In addition, this article constructs a hybrid quantum convolutional neural network nameded HQCNN to extract the temporal features of ECG to detect abnormal heartbeats. The simulation experimental results show that HQCNN achieves an average training and testing accuracy of 94.7% and 93.6%. And the detection stability is much higher than classical CNN with the same structure. HQCNN also has certain robustness under the perturbation of quantum noise. Besides, this article demonstrates through mathematical analysis that the proposed quantum blockchain algorithm has strong security and can effectively resist various quantum attacks, such as external attacks, Entanglement-Measure attack and Interception-Measurement-Repeat attack.

15.
Artigo em Inglês | MEDLINE | ID: mdl-37368802

RESUMO

Nanorobots have been used in smart health to collect time series data such as electrocardiograms and electroencephalograms. Real-time classification of dynamic time series signals in nanorobots is a challenging task. Nanorobots in the nanoscale range require a classification algorithm with low computational complexity. First, the classification algorithm should be able to dynamically analyze time series signals and update itself to process the concept drifts (CD). Second, the classification algorithm should have the ability to handle catastrophic forgetting (CF) and classify historical data. Most importantly, the classification algorithm should be energy-efficient to use less computing power and memory to classify signals in real-time on a smart nanorobot. To solve these challenges, we design an algorithm that can Prevent Concept Drift in Online continual Learning for time series classification (PCDOL). The prototype suppression item in PCDOL can reduce the impact caused by CD. It also solves the CF problem through the replay feature. The computation per second and the memory consumed by PCDOL are only 3.572M and 1KB, respectively. The experimental results show that PCDOL is better than several state-of-the-art methods for dealing with CD and CF in energy-efficient nanorobots.

16.
Stud Health Technol Inform ; 302: 603-604, 2023 May 18.
Artigo em Inglês | MEDLINE | ID: mdl-37203757

RESUMO

Understanding the aspects of progression for atherosclerotic cardiovascular disease and treatment is key to building reliable clinical decision-support systems. To promote system trust, one step is to make the machine learning models (used by the decision support systems) explainable for clinicians, developers, and researchers. Recently, working with longitudinal clinical trajectories using Graph Neural Networks (GNNs) has attracted attention among machine learning researchers. Although GNNs are seen as black-box methods, promising explainable AI (XAI) methods for GNNs have lately been proposed. In this paper, which describes initial project stages, we aim at utilizing GNNs for modeling, predicting, and exploring the model explainability of the low-density lipoprotein cholesterol level in long-term atherosclerotic cardiovascular disease progression and treatment.


Assuntos
Doenças Cardiovasculares , Sistemas de Apoio a Decisões Clínicas , Humanos , Aprendizado de Máquina , Redes Neurais de Computação , Pesquisadores
17.
Artigo em Inglês | MEDLINE | ID: mdl-37027677

RESUMO

Continuously analyzing medical time series as new classes emerge is meaningful for health monitoring and medical decision-making. Few-shot class-incremental learning (FSCIL) explores the classification of few-shot new classes without forgetting old classes. However, little of the existing research on FSCIL focuses on medical time series classification, which is more challenging to learn due to its large intra-class variability. In this paper, we propose a framework, the Meta self-Attention Prototype Incrementer (MAPIC) to address these problems. MAPIC contains three main modules: an embedding encoder for feature extraction, a prototype enhancement module for increasing inter-class variation, and a distance-based classifier for reducing intra-class variation. To mitigate catastrophic forgetting, MAPIC adopts a parameter protection strategy in which the parameters of the embedding encoder module are frozen at incremental stages after being trained in the base stage. The prototype enhancement module is proposed to enhance the expressiveness of prototypes by perceiving inter-class relations using a self-attention mechanism. We design a composite loss function containing the sample classification loss, the prototype non-overlapping loss, and the knowledge distillation loss, which work together to reduce intra-class variations and resist catastrophic forgetting. Experimental results on three different time series datasets show that MAPIC significantly outperforms state-of-the-art approaches by 27.99%, 18.4%, and 3.95%, respectively.

18.
Sci Rep ; 13(1): 4124, 2023 03 13.
Artigo em Inglês | MEDLINE | ID: mdl-36914679

RESUMO

Industrial Internet of Things (IIoT) is the new paradigm to perform different healthcare  applications with different services in daily life. Healthcare applications based on IIoT paradigm are widely used to track patients health status using remote healthcare technologies. Complex biomedical sensors exploit wireless technologies, and remote services in terms of industrial workflow applications to perform different healthcare tasks, such as like heartbeat, blood pressure and others. However, existing industrial healthcare technoloiges still has to deal with many problems, such as security, task scheduling, and the cost of processing tasks in IIoT based healthcare paradigms. This paper proposes a new solution to the above-mentioned issues and presents the deep reinforcement learning-aware blockchain-based task scheduling (DRLBTS) algorithm framework with different goals. DRLBTS provides security and makespan efficient scheduling for the healthcare applications. Then, it shares secure and valid data between connected network nodes after the initial assignment and data validation. Statistical results show that DRLBTS is adaptive and meets the security, privacy, and makespan requirements of healthcare applications in the distributed network.


Assuntos
Blockchain , Humanos , Algoritmos , Conscientização , Tecnologia Biomédica , Atenção à Saúde , Segurança Computacional
19.
Neural Comput Appl ; 35(19): 13907-13920, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-34127892

RESUMO

Major countries are globally facing difficult situations due to this pandemic disease, COVID-19. There are high chances of getting false positives and false negatives identifying the COVID-19 symptoms through existing medical practices such as PCR (polymerase chain reaction) and RT-PCR (reverse transcription-polymerase chain reaction). It might lead to a community spread of the disease. The alternative of these tests can be CT (Computer Tomography) imaging or X-rays of the lungs to identify the patient with COVID-19 symptoms more accurately. Furthermore, by using feasible and usable technology to automate the identification of COVID-19, the facilities can be improved. This notion became the basic framework, Res-CovNet, of the implemented methodology, a hybrid methodology to bring different platforms into a single platform. This basic framework is incorporated into IoMT based framework, a web-based service to identify and classify various forms of pneumonia or COVID-19 utilizing chest X-ray images. For the front end, the.NET framework along with C# language was utilized, MongoDB was utilized for the storage aspect, Res-CovNet was utilized for the processing aspect. Deep learning combined with the notion forms a comprehensive implementation of the framework, Res-CovNet, to classify the COVID-19 affected patients from pneumonia-affected patients as both lung imaging looks similar to the naked eye. The implemented framework, Res-CovNet, developed with the technique, transfer learning in which ResNet-50 used as a pre-trained model and then extended with classification layers. The work implemented using the data of X-ray images collected from the various trustable sources that include cases such as normal, bacterial pneumonia, viral pneumonia, and COVID-19, with the overall size of the data is about 5856. The accuracy of the model implemented is about 98.4% in identifying COVID-19 against the normal cases. The accuracy of the model is about 96.2% in the case of identifying COVID-19 against all other cases, as mentioned.

20.
IEEE J Biomed Health Inform ; 27(2): 823-834, 2023 02.
Artigo em Inglês | MEDLINE | ID: mdl-35041615

RESUMO

Internet of medical things (IoMT) has made it possible to collect applications and medical devices to improve healthcare information technology. Since the advent of the pandemic of coronavirus (COVID-19) in 2019, public health information has become more sensitive than ever. Moreover, different news items incorporated have resulted in differing public perceptions of COVID-19, especially on the social media platform and infrastructure. In addition, the unprecedented virality and changing nature of COVID-19 makes call centres to be likely overstressed, which is due to a lack of authentic and unregulated public media information. Furthermore, the lack of data privacy has restricted the sharing of COVID-19 information among health institutions. To resolve the above-mentioned limitations, this paper is proposing a privacy infrastructure based on federated learning and blockchain. The proposed infrastructure has the potentials to enhance the trust and authenticity of public media to disseminate COVID-19 information. Also, the proposed infrastructure can effectively provide a shared model while preserving the privacy of data owners. Furthermore, information security and privacy analyses show that the proposed infrastructure is robust against information security-related attacks.


Assuntos
Blockchain , COVID-19 , Humanos , Segurança Computacional , Atenção à Saúde , Privacidade
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA