Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 31.682
Filter
1.
Brief Bioinform ; 25(4)2024 May 23.
Article in English | MEDLINE | ID: mdl-38980373

ABSTRACT

Inferring gene regulatory networks (GRNs) allows us to obtain a deeper understanding of cellular function and disease pathogenesis. Recent advances in single-cell RNA sequencing (scRNA-seq) technology have improved the accuracy of GRN inference. However, many methods for inferring individual GRNs from scRNA-seq data are limited because they overlook intercellular heterogeneity and similarities between different cell subpopulations, which are often present in the data. Here, we propose a deep learning-based framework, DeepGRNCS, for jointly inferring GRNs across cell subpopulations. We follow the commonly accepted hypothesis that the expression of a target gene can be predicted based on the expression of transcription factors (TFs) due to underlying regulatory relationships. We initially processed scRNA-seq data by discretizing data scattering using the equal-width method. Then, we trained deep learning models to predict target gene expression from TFs. By individually removing each TF from the expression matrix, we used pre-trained deep model predictions to infer regulatory relationships between TFs and genes, thereby constructing the GRN. Our method outperforms existing GRN inference methods for various simulated and real scRNA-seq datasets. Finally, we applied DeepGRNCS to non-small cell lung cancer scRNA-seq data to identify key genes in each cell subpopulation and analyzed their biological relevance. In conclusion, DeepGRNCS effectively predicts cell subpopulation-specific GRNs. The source code is available at https://github.com/Nastume777/DeepGRNCS.


Subject(s)
Deep Learning , Gene Regulatory Networks , Single-Cell Analysis , Humans , Single-Cell Analysis/methods , Transcription Factors/genetics , Transcription Factors/metabolism , Computational Biology/methods , Sequence Analysis, RNA/methods , RNA-Seq/methods
2.
Brief Bioinform ; 25(4)2024 May 23.
Article in English | MEDLINE | ID: mdl-38980372

ABSTRACT

Around 50 years ago, molecular biology opened the path to understand changes in forms, adaptations, complexity, or the basis of human diseases through myriads of reports on gene birth, gene duplication, gene expression regulation, and splicing regulation, among other relevant mechanisms behind gene function. Here, with the advent of big data and artificial intelligence (AI), we focus on an elusive and intriguing mechanism of gene function regulation, RNA editing, in which a single nucleotide from an RNA molecule is changed, with a remarkable impact in the increase of the complexity of the transcriptome and proteome. We present a new generation approach to assess the functional conservation of the RNA-editing targeting mechanism using two AI learning algorithms, random forest (RF) and bidirectional long short-term memory (biLSTM) neural networks with an attention layer. These algorithms, combined with RNA-editing data coming from databases and variant calling from same-individual RNA and DNA-seq experiments from different species, allowed us to predict RNA-editing events using both primary sequence and secondary structure. Then, we devised a method for assessing conservation or divergence in the molecular mechanisms of editing completely in silico: the cross-testing analysis. This novel method not only helps to understand the conservation of the editing mechanism through evolution but could set the basis for achieving a better understanding of the adenosine-targeting mechanism in other fields.


Subject(s)
Machine Learning , RNA Editing , Humans , Algorithms , Computer Simulation , Computational Biology/methods , Neural Networks, Computer , RNA/genetics , RNA/metabolism
3.
Brief Bioinform ; 25(4)2024 May 23.
Article in English | MEDLINE | ID: mdl-38980375

ABSTRACT

Structural variation (SV) is an important form of genomic variation that influences gene function and expression by altering the structure of the genome. Although long-read data have been proven to better characterize SVs, SVs detected from noisy long-read data still include a considerable portion of false-positive calls. To accurately detect SVs in long-read data, we present SVDF, a method that employs a learning-based noise filtering strategy and an SV signature-adaptive clustering algorithm, for effectively reducing the likelihood of false-positive events. Benchmarking results from multiple orthogonal experiments demonstrate that, across different sequencing platforms and depths, SVDF achieves higher calling accuracy for each sample compared to several existing general SV calling tools. We believe that, with its meticulous and sensitive SV detection capability, SVDF can bring new opportunities and advancements to cutting-edge genomic research.


Subject(s)
Algorithms , Humans , Sequence Analysis, DNA/methods , High-Throughput Nucleotide Sequencing/methods , Genomics/methods , Genomic Structural Variation , Software
4.
J Imaging Inform Med ; 2024 Jul 09.
Article in English | MEDLINE | ID: mdl-38980623

ABSTRACT

Malposition of a nasogastric tube (NGT) can lead to severe complications. We aimed to develop a computer-aided detection (CAD) system to localize NGTs and detect NGT malposition on portable chest X-rays (CXRs). A total of 7378 portable CXRs were retrospectively retrieved from two hospitals between 2015 and 2020. All CXRs were annotated with pixel-level labels for NGT localization and image-level labels for NGT presence and malposition. In the CAD system, DeepLabv3 + with backbone ResNeSt50 and DenseNet121 served as the model architecture for segmentation and classification models, respectively. The CAD system was tested on images from chronologically different datasets (National Taiwan University Hospital (National Taiwan University Hospital)-20), geographically different datasets (National Taiwan University Hospital-Yunlin Branch (YB)), and the public CLiP dataset. For the segmentation model, the Dice coefficients indicated accurate delineation of the NGT course (National Taiwan University Hospital-20: 0.665, 95% confidence interval (CI) 0.630-0.696; National Taiwan University Hospital-Yunlin Branch: 0.646, 95% CI 0.614-0.678). The distance between the predicted and ground-truth NGT tips suggested accurate tip localization (National Taiwan University Hospital-20: 1.64 cm, 95% CI 0.99-2.41; National Taiwan University Hospital-Yunlin Branch: 2.83 cm, 95% CI 1.94-3.76). For the classification model, NGT presence was detected with high accuracy (area under the receiver operating characteristic curve (AUC): National Taiwan University Hospital-20: 0.998, 95% CI 0.995-1.000; National Taiwan University Hospital-Yunlin Branch: 0.998, 95% CI 0.995-1.000; CLiP dataset: 0.991, 95% CI 0.990-0.992). The CAD system also detected NGT malposition with high accuracy (AUC: National Taiwan University Hospital-20: 0.964, 95% CI 0.917-1.000; National Taiwan University Hospital-Yunlin Branch: 0.991, 95% CI 0.970-1.000) and detected abnormal nasoenteric tube positions with favorable performance (AUC: 0.839, 95% CI 0.807-0.869). The CAD system accurately localized NGTs and detected NGT malposition, demonstrating excellent potential for external generalizability.

5.
J Imaging Inform Med ; 2024 Jul 09.
Article in English | MEDLINE | ID: mdl-38980627

ABSTRACT

Accurate image classification and retrieval are of importance for clinical diagnosis and treatment decision-making. The recent contrastive language-image pre-training (CLIP) model has shown remarkable proficiency in understanding natural images. Drawing inspiration from CLIP, pathology-dedicated CLIP (PathCLIP) has been developed, utilizing over 200,000 image and text pairs in training. While the performance the PathCLIP is impressive, its robustness under a wide range of image corruptions remains unknown. Therefore, we conduct an extensive evaluation to analyze the performance of PathCLIP on various corrupted images from the datasets of osteosarcoma and WSSS4LUAD. In our experiments, we introduce eleven corruption types including brightness, contrast, defocus, resolution, saturation, hue, markup, deformation, incompleteness, rotation, and flipping at various settings. Through experiments, we find that PathCLIP surpasses OpenAI-CLIP and the pathology language-image pre-training (PLIP) model in zero-shot classification. It is relatively robust to image corruptions including contrast, saturation, incompleteness, and orientation factors. Among the eleven corruptions, hue, markup, deformation, defocus, and resolution can cause relatively severe performance fluctuation of the PathCLIP. This indicates that ensuring the quality of images is crucial before conducting a clinical test. Additionally, we assess the robustness of PathCLIP in the task of image-to-image retrieval, revealing that PathCLIP performs less effectively than PLIP on osteosarcoma but performs better on WSSS4LUAD under diverse corruptions. Overall, PathCLIP presents impressive zero-shot classification and retrieval performance for pathology images, but appropriate care needs to be taken when using it.

6.
Article in English | MEDLINE | ID: mdl-38976178

ABSTRACT

BACKGROUND AND OBJECTIVE: Accurate IVD segmentation is crucial for diagnosing and treating spinal conditions. Traditional deep learning methods depend on extensive, annotated datasets, which are hard to acquire. This research proposes an intensity-based self-supervised domain adaptation, using unlabeled multi-domain data to reduce reliance on large annotated datasets. METHODS: The study introduces an innovative method using intensity-based self-supervised learning for IVD segmentation in MRI scans. This approach is particularly suited for IVD segmentations due to its ability to effectively capture the subtle intensity variations that are characteristic of spinal structures. The model, a dual-task system, simultaneously segments IVDs and predicts intensity transformations. This intensity-focused method has the advantages of being easy to train and computationally light, making it highly practical in diverse clinical settings. Trained on unlabeled data from multiple domains, the model learns domain-invariant features, adeptly handling intensity variations across different MRI devices and protocols. RESULTS: Testing on three public datasets showed that this model outperforms baseline models trained on single-domain data. It handles domain shifts and achieves higher accuracy in IVD segmentation. CONCLUSIONS: This study demonstrates the potential of intensity-based self-supervised domain adaptation for IVD segmentation. It suggests new directions for research in enhancing generalizability across datasets with domain shifts, which can be applied to other medical imaging fields.

7.
Se Pu ; 42(7): 669-680, 2024 Jul.
Article in Chinese | MEDLINE | ID: mdl-38966975

ABSTRACT

Mass spectrometry imaging (MSI) is a promising method for characterizing the spatial distribution of compounds. Given the diversified development of acquisition methods and continuous improvements in the sensitivity of this technology, both the total amount of generated data and complexity of analysis have exponentially increased, rendering increasing challenges of data postprocessing, such as large amounts of noise, background signal interferences, as well as image registration deviations caused by sample position changes and scan deviations, and etc. Deep learning (DL) is a powerful tool widely used in data analysis and image reconstruction. This tool enables the automatic feature extraction of data by building and training a neural network model, and achieves comprehensive and in-depth analysis of target data through transfer learning, which has great potential for MSI data analysis. This paper reviews the current research status, application progress and challenges of DL in MSI data analysis, focusing on four core stages: data preprocessing, image reconstruction, cluster analysis, and multimodal fusion. The application of a combination of DL and mass spectrometry imaging in the study of tumor diagnosis and subtype classification is also illustrated. This review also discusses trends of development in the future, aiming to promote a better combination of artificial intelligence and mass spectrometry technology.


Subject(s)
Deep Learning , Image Processing, Computer-Assisted , Mass Spectrometry , Mass Spectrometry/methods , Image Processing, Computer-Assisted/methods , Humans , Data Analysis
8.
J Cancer Res Clin Oncol ; 150(7): 346, 2024 Jul 09.
Article in English | MEDLINE | ID: mdl-38981916

ABSTRACT

PURPOSE: To develop a deep learning (DL) model for differentiating between benign and malignant ovarian tumors of Ovarian-Adnexal Reporting and Data System Ultrasound (O-RADS US) Category 4 lesions, and validate its diagnostic performance. METHODS: A retrospective analysis of 1619 US images obtained from three centers from December 2014 to March 2023. DeepLabV3 and YOLOv8 were jointly used to segment, classify, and detect ovarian tumors. Precision and recall and area under the receiver operating characteristic curve (AUC) were employed to assess the model performance. RESULTS: A total of 519 patients (including 269 benign and 250 malignant masses) were enrolled in the study. The number of women included in the training, validation, and test cohorts was 426, 46, and 47, respectively. The detection models exhibited an average precision of 98.68% (95% CI: 0.95-0.99) for benign masses and 96.23% (95% CI: 0.92-0.98) for malignant masses. Moreover, in the training set, the AUC was 0.96 (95% CI: 0.94-0.97), whereas in the validation set, the AUC was 0.93(95% CI: 0.89-0.94) and 0.95 (95% CI: 0.91-0.96) in the test set. The sensitivity, specificity, accuracy, positive predictive value, and negative predictive values for the training set were 0.943,0.957,0.951,0.966, and 0.936, respectively, whereas those for the validation set were 0.905,0.935, 0.935,0.919, and 0.931, respectively. In addition, the sensitivity, specificity, accuracy, positive predictive value, and negative predictive value for the test set were 0.925, 0.955, 0.941, 0.956, and 0.927, respectively. CONCLUSION: The constructed DL model exhibited high diagnostic performance in distinguishing benign and malignant ovarian tumors in O-RADS US category 4 lesions.


Subject(s)
Deep Learning , Ovarian Neoplasms , Ultrasonography , Humans , Female , Ovarian Neoplasms/diagnostic imaging , Ovarian Neoplasms/pathology , Ovarian Neoplasms/diagnosis , Retrospective Studies , Middle Aged , Ultrasonography/methods , Adult , Aged , Young Adult
9.
Int Urol Nephrol ; 2024 Jul 09.
Article in English | MEDLINE | ID: mdl-38982018

ABSTRACT

BACKGROUND: Artificial intelligence (AI) has emerged as a promising avenue for improving patient care and surgical outcomes in urological surgery. However, the extent of AI's impact in predicting and managing complications is not fully elucidated. OBJECTIVES: We review the application of AI to foresee and manage complications in urological surgery, assess its efficacy, and discuss challenges to its use. METHODS AND MATERIALS: A targeted non-systematic literature search was conducted using the PubMed and Google Scholar databases to identify studies on AI in urological surgery and its complications. Evidence from the studies was synthesised. RESULTS: Incorporating AI into various facets of urological surgery has shown promising advancements. From preoperative planning to intraoperative guidance, AI is revolutionising the field, demonstrating remarkable proficiency in tasks such as image analysis, decision-making support, and complication prediction. Studies show that AI programmes are highly accurate, increase surgical precision and efficiency, and reduce complications. However, implementation challenges exist in AI errors, human errors, and ethical issues. CONCLUSION: AI has great potential in predicting and managing surgical complications of urological surgery. Advancements have been made, but challenges and ethical considerations must be addressed before widespread AI implementation.

10.
Endocrine ; 2024 Jul 09.
Article in English | MEDLINE | ID: mdl-38982023

ABSTRACT

BACKGROUND: It was essential to identify individuals at high risk of fragility fracture and prevented them due to the significant morbidity, mortality, and economic burden associated with fragility fracture. The quantitative ultrasound (QUS) showed promise in assessing bone structure characteristics and determining the risk of fragility fracture. AIMS: To evaluate the performance of a multi-channel residual network (MResNet) based on ultrasonic radiofrequency (RF) signal to discriminate fragility fractures retrospectively in postmenopausal women, and compared it with the traditional parameter of QUS, speed of sound (SOS), and bone mineral density (BMD) acquired with dual X-ray absorptiometry (DXA). METHODS: Using QUS, RF signal and SOS were acquired for 246 postmenopausal women. An MResNet was utilized, based on the RF signal, to categorize individuals with an elevated risk of fragility fracture. DXA was employed to obtain BMD at the lumbar, hip, and femoral neck. The fracture history of all adult subjects was gathered. Analyzing the odds ratios (OR) and the area under the receiver operator characteristic curves (AUC) was done to evaluate the effectiveness of various methods in discriminating fragility fracture. RESULTS: Among the 246 postmenopausal women, 170 belonged to the non-fracture group, 50 to the vertebral group, and 26 to the non-vertebral fracture group. MResNet was competent to discriminate any fragility fracture (OR = 2.64; AUC = 0.74), Vertebral fracture (OR = 3.02; AUC = 0.77), and non-vertebral fracture (OR = 2.01; AUC = 0.69). After being modified by clinical covariates, the efficiency of MResNet was further improved to OR = 3.31-4.08, AUC = 0.81-0.83 among all fracture groups, which significantly surpassed QUS-SOS (OR = 1.32-1.36; AUC = 0.60) and DXA-BMD (OR = 1.23-2.94; AUC = 0.63-0.76). CONCLUSIONS: This pilot cross-sectional study demonstrates that the MResNet model based on the ultrasonic RF signal shows promising performance in discriminating fragility fractures in postmenopausal women. When incorporating clinical covariates, the efficiency of the modified MResNet is further enhanced, surpassing the performance of QUS-SOS and DXA-BMD in terms of OR and AUC. These findings highlight the potential of the MResNet as a promising approach for fracture risk assessment. Future research should focus on larger and more diverse populations to validate these results and explore its clinical applications.

11.
Sci Rep ; 14(1): 15775, 2024 Jul 09.
Article in English | MEDLINE | ID: mdl-38982238

ABSTRACT

A three-dimensional convolutional neural network model was developed to classify the severity of chronic kidney disease (CKD) using magnetic resonance imaging (MRI) Dixon-based T1-weighted in-phase (IP)/opposed-phase (OP)/water-only (WO) imaging. Seventy-three patients with severe renal dysfunction (estimated glomerular filtration rate [eGFR] < 30 mL/min/1.73 m2, CKD stage G4-5); 172 with moderate renal dysfunction (30 ≤ eGFR < 60 mL/min/1.73 m2, CKD stage G3a/b); and 76 with mild renal dysfunction (eGFR ≥ 60 mL/min/1.73 m2, CKD stage G1-2) participated in this study. The model was applied to the right, left, and both kidneys, as well as to each imaging method (T1-weighted IP/OP/WO images). The best performance was obtained when using bilateral kidneys and IP images, with an accuracy of 0.862 ± 0.036. The overall accuracy was better for the bilateral kidney models than for the unilateral kidney models. Our deep learning approach using kidney MRI can be applied to classify patients with CKD based on the severity of kidney disease.


Subject(s)
Glomerular Filtration Rate , Kidney , Magnetic Resonance Imaging , Neural Networks, Computer , Renal Insufficiency, Chronic , Severity of Illness Index , Humans , Renal Insufficiency, Chronic/diagnostic imaging , Renal Insufficiency, Chronic/pathology , Magnetic Resonance Imaging/methods , Female , Male , Middle Aged , Kidney/diagnostic imaging , Kidney/pathology , Aged , Adult , Deep Learning , Imaging, Three-Dimensional/methods
12.
BMC Med Imaging ; 24(1): 170, 2024 Jul 09.
Article in English | MEDLINE | ID: mdl-38982357

ABSTRACT

OBJECTIVES: To develop and validate a novel interpretable artificial intelligence (AI) model that integrates radiomic features, deep learning features, and imaging features at multiple semantic levels to predict the prognosis of intracerebral hemorrhage (ICH) patients at 6 months post-onset. MATERIALS AND METHODS: Retrospectively enrolled 222 patients with ICH for Non-contrast Computed Tomography (NCCT) images and clinical data, who were divided into a training cohort (n = 186, medical center 1) and an external testing cohort (n = 36, medical center 2). Following image preprocessing, the entire hematoma region was segmented by two radiologists as the volume of interest (VOI). Pyradiomics algorithm library was utilized to extract 1762 radiomics features, while a deep convolutional neural network (EfficientnetV2-L) was employed to extract 1000 deep learning features. Additionally, radiologists evaluated imaging features. Based on the three different modalities of features mentioned above, the Random Forest (RF) model was trained, resulting in three models (Radiomics Model, Radiomics-Clinical Model, and DL-Radiomics-Clinical Model). The performance and clinical utility of the models were assessed using the Area Under the Receiver Operating Characteristic Curve (AUC), calibration curve, and Decision Curve Analysis (DCA), with AUC compared using the DeLong test. Furthermore, this study employs three methods, Shapley Additive Explanations (SHAP), Grad-CAM, and Guided Grad-CAM, to conduct a multidimensional interpretability analysis of model decisions. RESULTS: The Radiomics-Clinical Model and DL-Radiomics-Clinical Model exhibited relatively good predictive performance, with an AUC of 0.86 [95% Confidence Intervals (CI): 0.71, 0.95; P < 0.01] and 0.89 (95% CI: 0.74, 0.97; P < 0.01), respectively, in the external testing cohort. CONCLUSION: The multimodal explainable AI model proposed in this study can accurately predict the prognosis of ICH. Interpretability methods such as SHAP, Grad-CAM, and Guided Grad-Cam partially address the interpretability limitations of AI models. Integrating multimodal imaging features can effectively improve the performance of the model. CLINICAL RELEVANCE STATEMENT: Predicting the prognosis of patients with ICH is a key objective in emergency care. Accurate and efficient prognostic tools can effectively prevent, manage, and monitor adverse events in ICH patients, maximizing treatment outcomes.


Subject(s)
Artificial Intelligence , Cerebral Hemorrhage , Deep Learning , Tomography, X-Ray Computed , Humans , Cerebral Hemorrhage/diagnostic imaging , Prognosis , Tomography, X-Ray Computed/methods , Male , Female , Retrospective Studies , Middle Aged , Aged , ROC Curve , Neural Networks, Computer , Algorithms
13.
Environ Sci Technol ; 2024 Jul 09.
Article in English | MEDLINE | ID: mdl-38982681

ABSTRACT

Gaseous nitrous acid (HONO) is identified as a critical precursor of hydroxyl radicals (OH), influencing atmospheric oxidation capacity and the formation of secondary pollutants. However, large uncertainties persist regarding its formation and elimination mechanisms, impeding accurate simulation of HONO levels using chemical models. In this study, a deep neural network (DNN) model was established based on routine air quality data (O3, NO2, CO, and PM2.5) and meteorological parameters (temperature, relative humidity, solar zenith angle, and season) collected from four typical megacity clusters in China. The model exhibited robust performance on both the train sets [slope = 1.0, r2 = 0.94, root mean squared error (RMSE) = 0.29 ppbv] and two independent test sets (slope = 1.0, r2 = 0.79, and RMSE = 0.39 ppbv), demonstrated excellent capability in reproducing the spatiotemporal variations of HONO, and outperformed an observation-constrained box model incorporated with newly proposed HONO formation mechanisms. Nitrogen dioxide (NO2) was identified as the most impactful features for HONO prediction using the SHapely Additive exPlanation (SHAP) approach, highlighting the importance of NO2 conversion in HONO formation. The DNN model was further employed to predict the future change of HONO levels in different NOx abatement scenarios, which is expected to decrease 27-44% in summer as the result of 30-50% NOx reduction. These results suggest a dual effect brought by abatement of NOx emissions, leading to not only reduction of O3 and nitrate precursors but also decrease in HONO levels and hence primary radical production rates (PROx). In summary, this study demonstrates the feasibility of using deep learning approach to predict HONO concentrations, offering a promising supplement to traditional chemical models. Additionally, stringent NOx abatement would be beneficial for collaborative alleviation of O3 and secondary PM2.5.

14.
Schizophr Bull ; 2024 Jul 09.
Article in English | MEDLINE | ID: mdl-38982882

ABSTRACT

BACKGROUND AND HYPOTHESIS: Schizophrenia (SZ) is characterized by significant cognitive and behavioral disruptions. Neuroimaging techniques, particularly magnetic resonance imaging (MRI), have been widely utilized to investigate biomarkers of SZ, distinguish SZ from healthy conditions or other mental disorders, and explore biotypes within SZ or across SZ and other mental disorders, which aim to promote the accurate diagnosis of SZ. In China, research on SZ using MRI has grown considerably in recent years. STUDY DESIGN: The article reviews advanced neuroimaging and artificial intelligence (AI) methods using single-modal or multimodal MRI to reveal the mechanism of SZ and promote accurate diagnosis of SZ, with a particular emphasis on the achievements made by Chinese scholars around the past decade. STUDY RESULTS: Our article focuses on the methods for capturing subtle brain functional and structural properties from the high-dimensional MRI data, the multimodal fusion and feature selection methods for obtaining important and sparse neuroimaging features, the supervised statistical analysis and classification for distinguishing disorders, and the unsupervised clustering and semi-supervised learning methods for identifying neuroimage-based biotypes. Crucially, our article highlights the characteristics of each method and underscores the interconnections among various approaches regarding biomarker extraction and neuroimage-based diagnosis, which is beneficial not only for comprehending SZ but also for exploring other mental disorders. CONCLUSIONS: We offer a valuable review of advanced neuroimage analysis and AI methods primarily focused on SZ research by Chinese scholars, aiming to promote the diagnosis, treatment, and prevention of SZ, as well as other mental disorders, both within China and internationally.

15.
PeerJ Comput Sci ; 10: e2140, 2024.
Article in English | MEDLINE | ID: mdl-38983198

ABSTRACT

The hiding capacity of the current information hiding field has reached a relatively high level, which can hide two color images into one color image. In order to explore a larger hidden capacity, an information hiding scheme based on an improved FCdDNet is proposed, which can hide large-size color images into small-size color images. An improved FCdDNet network is used as the main structure shared by the hidden network and the extraction network. These two networks promote and improve each other during the confrontation training process and are used in pairs. It can be seen that the proposed scheme achieves a larger information hiding capacity, and the hidden information is four times larger than the scale of the carrier image. At the same time, the visual effect after hiding is guaranteed, and the image extracted from the hidden image also has a high degree of restoration. The scheme can be applied to image authentication, secret image transmission, and other fields.

16.
PeerJ Comput Sci ; 10: e2103, 2024.
Article in English | MEDLINE | ID: mdl-38983199

ABSTRACT

Images and videos containing fake faces are the most common type of digital manipulation. Such content can lead to negative consequences by spreading false information. The use of machine learning algorithms to produce fake face images has made it challenging to distinguish between genuine and fake content. Face manipulations are categorized into four basic groups: entire face synthesis, face identity manipulation (deepfake), facial attribute manipulation and facial expression manipulation. The study utilized lightweight convolutional neural networks to detect fake face images generated by using entire face synthesis and generative adversarial networks. The dataset used in the training process includes 70,000 real images in the FFHQ dataset and 70,000 fake images produced with StyleGAN2 using the FFHQ dataset. 80% of the dataset was used for training and 20% for testing. Initially, the MobileNet, MobileNetV2, EfficientNetB0, and NASNetMobile convolutional neural networks were trained separately for the training process. In the training, the models were pre-trained on ImageNet and reused with transfer learning. As a result of the first trainings EfficientNetB0 algorithm reached the highest accuracy of 93.64%. The EfficientNetB0 algorithm was revised to increase its accuracy rate by adding two dense layers (256 neurons) with ReLU activation, two dropout layers, one flattening layer, one dense layer (128 neurons) with ReLU activation function, and a softmax activation function used for the classification dense layer with two nodes. As a result of this process accuracy rate of 95.48% was achieved with EfficientNetB0 algorithm. Finally, the model that achieved 95.48% accuracy was used to train MobileNet and MobileNetV2 models together using the stacking ensemble learning method, resulting in the highest accuracy rate of 96.44%.

17.
PeerJ Comput Sci ; 10: e2125, 2024.
Article in English | MEDLINE | ID: mdl-38983197

ABSTRACT

This study proposes a novel hybrid model, called ICE2DE-MDL, integrating secondary decomposition, entropy, machine and deep learning methods to predict a stock closing price. In this context, first of all, the noise contained in the financial time series was eliminated. A denoising method, which utilizes entropy and the two-level ICEEMDAN methodology, is suggested to achieve this. Subsequently, we applied many deep learning and machine learning methods, including long-short term memory (LSTM), LSTM-BN, gated recurrent unit (GRU), and SVR, to the IMFs obtained from the decomposition, classifying them as noiseless. Afterward, the best training method was determined for each IMF. Finally, the proposed model's forecast was obtained by hierarchically combining the prediction results of each IMF. The ICE2DE-MDL model was applied to eight stock market indices and three stock data sets, and the next day's closing price of these stock items was predicted. The results indicate that RMSE values ranged from 0.031 to 0.244, MAE values ranged from 0.026 to 0.144, MAPE values ranged from 0.128 to 0.594, and R-squared values ranged from 0.905 to 0.998 for stock indices and stock forecasts. Furthermore, comparisons were made with various hybrid models proposed within the scope of stock forecasting to evaluate the performance of the ICE2DE-MDL model. Upon comparison, The ICE2DE-MDL model demonstrated superior performance relative to existing models in the literature for both forecasting stock market indices and individual stocks. Additionally, to our knowledge, this study is the first to effectively eliminate noise in stock item data using the concepts of entropy and ICEEMDAN. It is also the second study to apply ICEEMDAN to a financial time series prediction problem.

18.
PeerJ Comput Sci ; 10: e2088, 2024.
Article in English | MEDLINE | ID: mdl-38983229

ABSTRACT

Fraudulent activities especially in auto insurance and credit card transactions impose significant financial losses on businesses and individuals. To overcome this issue, we propose a novel approach for fraud detection, combining convolutional neural networks (CNNs) with support vector machine (SVM), k nearest neighbor (KNN), naive Bayes (NB), and decision tree (DT) algorithms. The core of this methodology lies in utilizing the deep features extracted from the CNNs as inputs to various machine learning models, thus significantly contributing to the enhancement of fraud detection accuracy and efficiency. Our results demonstrate superior performance compared to previous studies, highlighting our model's potential for widespread adoption in combating fraudulent activities.

19.
PeerJ Comput Sci ; 10: e2077, 2024.
Article in English | MEDLINE | ID: mdl-38983227

ABSTRACT

Background: Dyslexia is a neurological disorder that affects an individual's language processing abilities. Early care and intervention can help dyslexic individuals succeed academically and socially. Recent developments in deep learning (DL) approaches motivate researchers to build dyslexia detection models (DDMs). DL approaches facilitate the integration of multi-modality data. However, there are few multi-modality-based DDMs. Methods: In this study, the authors built a DL-based DDM using multi-modality data. A squeeze and excitation (SE) integrated MobileNet V3 model, self-attention mechanisms (SA) based EfficientNet B7 model, and early stopping and SA-based Bi-directional long short-term memory (Bi-LSTM) models were developed to extract features from magnetic resonance imaging (MRI), functional MRI, and electroencephalography (EEG) data. In addition, the authors fine-tuned the LightGBM model using the Hyperband optimization technique to detect dyslexia using the extracted features. Three datasets containing FMRI, MRI, and EEG data were used to evaluate the performance of the proposed DDM. Results: The findings supported the significance of the proposed DDM in detecting dyslexia with limited computational resources. The proposed model outperformed the existing DDMs by producing an optimal accuracy of 98.9%, 98.6%, and 98.8% for the FMRI, MRI, and EEG datasets, respectively. Healthcare centers and educational institutions can benefit from the proposed model to identify dyslexia in the initial stages. The interpretability of the proposed model can be improved by integrating vision transformers-based feature extraction.

20.
PeerJ Comput Sci ; 10: e2039, 2024.
Article in English | MEDLINE | ID: mdl-38983232

ABSTRACT

As more aerial imagery becomes readily available, massive volumes of data are being gathered constantly. Several groups can benefit from the data provided by this geographical imagery. However, it is time-consuming to manually analyze each image to gain information on land cover. This research suggests using deep learning methods for precise and rapid pixel-by-pixel classification of aerial imagery for land cover analysis, which would be a significant step forward in resolving this issue. The suggested method has several steps, such as the augmentation and transformation of data, the selection of deep learning models, and the final prediction. The study uses the three most popular deep learning models (Vanilla-UNet, ResNet50 UNet, and DeepLabV3 ResNet50) for the experiments. According to the experimental results, the ResNet50 UNet model achieved an accuracy of 94.37%, the DeepLabV3 ResNet50 model achieved an accuracy of 94.77%, and the Vanilla-UNet model achieved an accuracy of 91.31%. The accuracy, precision, recall, and F1-score of DeepLabV3 and ResNet50 are higher than those of the other two models. The proposed approach is also compared to the existing UNet approach, and the proposed approaches have produced greater probability prediction scores than the conventional UNet model for all classes. Our approach outperforms model DeepLabV3 ResNet50 on aerial image datasets based on the performance.

SELECTION OF CITATIONS
SEARCH DETAIL
...