Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 44
Filtrar
1.
BMC Bioinformatics ; 25(1): 108, 2024 Mar 12.
Artículo en Inglés | MEDLINE | ID: mdl-38475723

RESUMEN

RNA-protein interaction (RPI) is crucial to the life processes of diverse organisms. Various researchers have identified RPI through long-term and high-cost biological experiments. Although numerous machine learning and deep learning-based methods for predicting RPI currently exist, their robustness and generalizability have significant room for improvement. This study proposes LPI-MFF, an RPI prediction model based on multi-source information fusion, to address these issues. The LPI-MFF employed protein-protein interactions features, sequence features, secondary structure features, and physical and chemical properties as the information sources with the corresponding coding scheme, followed by the random forest algorithm for feature screening. Finally, all information was combined and a classification method based on convolutional neural networks is used. The experimental results of fivefold cross-validation demonstrated that the accuracy of LPI-MFF on RPI1807 and NPInter was 97.60% and 97.67%, respectively. In addition, the accuracy rate on the independent test set RPI1168 was 84.9%, and the accuracy rate on the Mus musculus dataset was 90.91%. Accordingly, LPI-MFF demonstrated greater robustness and generalization than other prevalent RPI prediction methods.


Asunto(s)
Aprendizaje Profundo , ARN Largo no Codificante , Animales , Ratones , ARN Largo no Codificante/química , Bosques Aleatorios , Redes Neurales de la Computación , Aprendizaje Automático , Biología Computacional/métodos
2.
Neuroimage ; 292: 120608, 2024 Apr 15.
Artículo en Inglés | MEDLINE | ID: mdl-38626817

RESUMEN

The morphological analysis and volume measurement of the hippocampus are crucial to the study of many brain diseases. Therefore, an accurate hippocampal segmentation method is beneficial for the development of clinical research in brain diseases. U-Net and its variants have become prevalent in hippocampus segmentation of Magnetic Resonance Imaging (MRI) due to their effectiveness, and the architecture based on Transformer has also received some attention. However, some existing methods focus too much on the shape and volume of the hippocampus rather than its spatial information, and the extracted information is independent of each other, ignoring the correlation between local and global features. In addition, many methods cannot be effectively applied to practical medical image segmentation due to many parameters and high computational complexity. To this end, we combined the advantages of CNNs and ViTs (Vision Transformer) and proposed a simple and lightweight model: Light3DHS for the segmentation of the 3D hippocampus. In order to obtain richer local contextual features, the encoder first utilizes a multi-scale convolutional attention module (MCA) to learn the spatial information of the hippocampus. Considering the importance of local features and global semantics for 3D segmentation, we used a lightweight ViT to learn high-level features of scale invariance and further fuse local-to-global representation. To evaluate the effectiveness of encoder feature representation, we designed three decoders of different complexity to generate segmentation maps. Experiments on three common hippocampal datasets demonstrate that the network achieves more accurate hippocampus segmentation with fewer parameters. Light3DHS performs better than other state-of-the-art algorithms.


Asunto(s)
Hipocampo , Imagenología Tridimensional , Imagen por Resonancia Magnética , Hipocampo/diagnóstico por imagen , Humanos , Imagen por Resonancia Magnética/métodos , Imagenología Tridimensional/métodos , Redes Neurales de la Computación , Aprendizaje Profundo , Algoritmos
3.
Network ; : 1-28, 2024 Jun 11.
Artículo en Inglés | MEDLINE | ID: mdl-38860460

RESUMEN

In this paper, Quaternion Fractional Order Meixner Moments-based Deep Siamese Domain Adaptation Convolutional Neural Network-based Big Data Analytical Technique is proposed for improving Cloud Data Security (DSDA-CNN-QFOMM-BD-CDS). The proposed methodology comprises six phases: data collection, transmission, pre-processing, storage, analysis, and security of data. Big data analysis methodologies start with the data collection phase. Deep Siamese domain adaptation convolutional Neural Network (DSDA-CNN) is applied to categorize the types of attacks in the cloud database during the data analysis process. During data security phase, Quaternion Fractional Order Meixner Moments (QFOMM) is employed to protect the cloud data for encryption with decryption. The proposed method is implemented in JAVA and assessed using performance metrics, including precision, sensitivity, accuracy, recall, specificity, f-measure, computational complexity information loss, compression ratio, throughput, encryption time, decryption time. The performance of the proposed method offers 23.31%, 15.64%, 18.89% better accuracy and 36.69%, 17.25%, 19.96% less information loss. When compared to existing methods like Fractional order discrete Tchebyshev encryption fostered big data analytical model to maximize the safety of cloud data depend on Enhanced Elman spike neural network (EESNN-FrDTM-BD-CDS), an innovative scheme architecture for safe authentication along data sharing in cloud enabled Big data Environment (LZMA-DBSCAN-BD-CDS).

4.
Int J Neurosci ; 133(5): 512-522, 2023 May.
Artículo en Inglés | MEDLINE | ID: mdl-34042552

RESUMEN

BACKGROUND: Moyamoya disease (MMD) is a serious intracranial cerebrovascular disease. Cerebral hemorrhage caused by MMD will bring life risk to patients. Therefore, MMD detection is of great significance in the prevention of cerebral hemorrhage. In order to improve the accuracy of digital subtraction angiography (DSA) in the diagnosis of ischemic MMD, in this paper, a deep network architecture combined with 3D convolutional neural network (3D CNN) and bidirectional convolutional gated recurrent unit (BiConvGRU) is proposed to learn the spatiotemporal features for ischemic MMD detection. METHODS: Firstly, 2D convolutional neural network (2D CNN) is utilized to extract spatial features for each frame of DSA. Secondly, the long-term spatiotemporal features of DSA sequence are extracted by BiConvGRU. Thirdly, the short-term spatiotemporal features of DSA are further extracted by 3D convolutional neural network (3D CNN). In addition, different features are extracted when gray images and optical flow images pass through the network, and multiple features are extracted by features fusion. Finally, the fused features are utilized to classify. RESULTS: The proposed method was quantitatively evaluated on a data sets of 630 cases. The experimental results showed a detection accuracy of 0.9788, sensitivity and specificity were 0.9780 and 0.9796, respectively, and area under curve (AUC) was 0.9856. Compared with other methods, we can get the highest accuracy and AUC. CONCLUSIONS: The experimental results show that the proposed method is stable and reliable for ischemic MMD detection, which provides an option for doctors to accurately diagnose ischemic MMD.


Asunto(s)
Enfermedad de Moyamoya , Humanos , Enfermedad de Moyamoya/diagnóstico por imagen , Angiografía de Substracción Digital/métodos , Redes Neurales de la Computación , Hemorragia Cerebral
5.
Sensors (Basel) ; 23(15)2023 Jul 31.
Artículo en Inglés | MEDLINE | ID: mdl-37571614

RESUMEN

Foreign bodies often cause belt scratching and tearing, coal stacking, and plugging during the transportation of coal via belt conveyors. To overcome the problems of large parameters, heavy computational complexity, low classification accuracy, and poor processing speed in current classification networks, a novel network based on ESCBAM and multichannel feature fusion is proposed in this paper. Firstly, to improve the utilization rate of features and the network's ability to learn detailed information, a multi-channel feature fusion strategy was designed to fully integrate the independent feature information between each channel. Then, to reduce the computational amount while maintaining excellent feature extraction capability, an information fusion network was constructed, which adopted the depthwise separable convolution and improved residual network structure as the basic feature extraction unit. Finally, to enhance the understanding ability of image context and improve the feature performance of the network, a novel ESCBAM attention mechanism with strong generalization and portability was constructed by integrating space and channel features. The experimental results demonstrate that the proposed method has the advantages of fewer parameters, low computational complexity, high accuracy, and fast processing speed, which can effectively classify foreign bodies on the belt conveyor.

6.
J Biomed Inform ; 115: 103693, 2021 03.
Artículo en Inglés | MEDLINE | ID: mdl-33540076

RESUMEN

BACKGROUND: Diabetics has become a serious public health burden in China. Multiple complications appear with the progression of diabetics pose a serious threat to the quality of human life and health. We can prevent the progression of prediabetics to diabetics and delay the progression to diabetics by early identification of diabetics and prediabetics and timely intervention, which have positive significance for improving public health. OBJECTIVE: Using machine learning techniques, we establish the noninvasive diabetics risk prediction model based on tongue features fusion and predict the risk of prediabetics and diabetics. METHODS: Applying the type TFDA-1 Tongue Diagnosis Instrument, we collect tongue images, extract tongue features including color and texture features using TDAS, and extract the advanced tongue features with ResNet-50, achieve the fusion of the two features with GA_XGBT, finally establish the noninvasive diabetics risk prediction model and evaluate the performance of testing effectiveness. RESULTS: Cross-validation suggests the best performance of GA_XGBT model with fusion features, whose average CA is 0.821, the average AUROC is 0.924, the average AUPRC is 0.856, the average Precision is 0.834, the average Recall is 0.822, the average F1-score is 0.813. Test set suggests the best testing performance of GA_XGBT model, whose average CA is 0.81, the average AUROC is 0.918, the average AUPRC is 0.839, the average Precision is 0.821, the average Recall is 0.81, the average F1-score is 0.796. When we test prediabetics with GA_XGBT model, we find that the AUROC is 0.914, the Precision is 0.69, the Recall is 0.952, the F1-score is 0.8. When we test diabetics with GA_XGBT model, we find that the AUROC is 0.984, the Precision is 0.929, the Recall is 0.951, the F1-score is 0.94. CONCLUSIONS: Based on tongue features, the study uses classical machine learning algorithm and deep learning algorithm to maximum the respective advantages. We combine the prior knowledge and potential features together, establish the noninvasive diabetics risk prediction model with features fusion algorithm, and detect prediabetics and diabetics noninvasively. Our study presents a feasible method for establishing the association between diabetics and the tongue image information and prove that tongue image information is a potential marker which facilitates effective early diagnosis of prediabetics and diabetics.


Asunto(s)
Diabetes Mellitus , Estado Prediabético , China , Diabetes Mellitus/diagnóstico , Humanos , Aprendizaje Automático , Estado Prediabético/diagnóstico , Lengua
7.
Sensors (Basel) ; 21(23)2021 Nov 28.
Artículo en Inglés | MEDLINE | ID: mdl-34883944

RESUMEN

Human action recognition (HAR) has gained significant attention recently as it can be adopted for a smart surveillance system in Multimedia. However, HAR is a challenging task because of the variety of human actions in daily life. Various solutions based on computer vision (CV) have been proposed in the literature which did not prove to be successful due to large video sequences which need to be processed in surveillance systems. The problem exacerbates in the presence of multi-view cameras. Recently, the development of deep learning (DL)-based systems has shown significant success for HAR even for multi-view camera systems. In this research work, a DL-based design is proposed for HAR. The proposed design consists of multiple steps including feature mapping, feature fusion and feature selection. For the initial feature mapping step, two pre-trained models are considered, such as DenseNet201 and InceptionV3. Later, the extracted deep features are fused using the Serial based Extended (SbE) approach. Later on, the best features are selected using Kurtosis-controlled Weighted KNN. The selected features are classified using several supervised learning algorithms. To show the efficacy of the proposed design, we used several datasets, such as KTH, IXMAS, WVU, and Hollywood. Experimental results showed that the proposed design achieved accuracies of 99.3%, 97.4%, 99.8%, and 99.9%, respectively, on these datasets. Furthermore, the feature selection step performed better in terms of computational time compared with the state-of-the-art.


Asunto(s)
Aprendizaje Profundo , Algoritmos , Actividades Humanas , Humanos , Reconocimiento de Normas Patrones Automatizadas
8.
Sensors (Basel) ; 21(22)2021 Nov 15.
Artículo en Inglés | MEDLINE | ID: mdl-34833658

RESUMEN

Human Gait Recognition (HGR) is a biometric technique that has been utilized for security purposes for the last decade. The performance of gait recognition can be influenced by various factors such as wearing clothes, carrying a bag, and the walking surfaces. Furthermore, identification from differing views is a significant difficulty in HGR. Many techniques have been introduced in the literature for HGR using conventional and deep learning techniques. However, the traditional methods are not suitable for large datasets. Therefore, a new framework is proposed for human gait recognition using deep learning and best feature selection. The proposed framework includes data augmentation, feature extraction, feature selection, feature fusion, and classification. In the augmentation step, three flip operations were used. In the feature extraction step, two pre-trained models were employed, Inception-ResNet-V2 and NASNet Mobile. Both models were fine-tuned and trained using transfer learning on the CASIA B gait dataset. The features of the selected deep models were optimized using a modified three-step whale optimization algorithm and the best features were chosen. The selected best features were fused using the modified mean absolute deviation extended serial fusion (MDeSF) approach. Then, the final classification was performed using several classification algorithms. The experimental process was conducted on the entire CASIA B dataset and achieved an average accuracy of 89.0. Comparison with existing techniques showed an improvement in accuracy, recall rate, and computational time.


Asunto(s)
Aprendizaje Profundo , Algoritmos , Marcha , Humanos
9.
Sensors (Basel) ; 21(16)2021 Aug 12.
Artículo en Inglés | MEDLINE | ID: mdl-34450894

RESUMEN

Emotion recognition is an important research field for human-computer interaction. Audio-video emotion recognition is now attacked with deep neural network modeling tools. In published papers, as a rule, the authors show only cases of the superiority in multi-modality over audio-only or video-only modality. However, there are cases of superiority in uni-modality that can be found. In our research, we hypothesize that for fuzzy categories of emotional events, the within-modal and inter-modal noisy information represented indirectly in the parameters of the modeling neural network impedes better performance in the existing late fusion and end-to-end multi-modal network training strategies. To take advantage of and overcome the deficiencies in both solutions, we define a multi-modal residual perceptron network which performs end-to-end learning from multi-modal network branches, generalizing better multi-modal feature representation. For the proposed multi-modal residual perceptron network and the novel time augmentation for streaming digital movies, the state-of-the-art average recognition rate was improved to 91.4% for the Ryerson Audio-Visual Database of Emotional Speech and Song dataset and to 83.15% for the Crowd-Sourced Emotional Multi Modal Actors dataset. Moreover, the multi-modal residual perceptron network concept shows its potential for multi-modal applications dealing with signal sources not only of optical and acoustical types.


Asunto(s)
Emociones , Redes Neurales de la Computación , Humanos , Películas Cinematográficas , Habla
10.
Comput Electr Eng ; 90: 106960, 2021 Mar.
Artículo en Inglés | MEDLINE | ID: mdl-33518824

RESUMEN

In this work, we propose a deep learning framework for the classification of COVID-19 pneumonia infection from normal chest CT scans. In this regard, a 15-layered convolutional neural network architecture is developed which extracts deep features from the selected image samples - collected from the Radiopeadia. Deep features are collected from two different layers, global average pool and fully connected layers, which are later combined using the max-layer detail (MLD) approach. Subsequently, a Correntropy technique is embedded in the main design to select the most discriminant features from the pool of features. One-class kernel extreme learning machine classifier is utilized for the final classification to achieving an average accuracy of 95.1%, and the sensitivity, specificity & precision rate of 95.1%, 95%, & 94% respectively. To further verify our claims, detailed statistical analyses based on standard error mean (SEM) is also provided, which proves the effectiveness of our proposed prediction design.

11.
Sensors (Basel) ; 20(20)2020 Oct 17.
Artículo en Inglés | MEDLINE | ID: mdl-33080880

RESUMEN

Handwritten character recognition is increasingly important in a variety of automation fields, for example, authentication of bank signatures, identification of ZIP codes on letter addresses, and forensic evidence. Despite improved object recognition technologies, Pashto's hand-written character recognition (PHCR) remains largely unsolved due to the presence of many enigmatic hand-written characters, enormously cursive Pashto characters, and lack of research attention. We propose a convolutional neural network (CNN) model for recognition of Pashto hand-written characters for the first time in an unrestricted environment. Firstly, a novel Pashto handwritten character data set, "Poha", for 44 characters is constructed. For preprocessing, deep fusion image processing techniques and noise reduction for text optimization are applied. A CNN model optimized in the number of convolutional layers and their parameters outperformed common deep models in terms of accuracy. Moreover, a set of benchmark popular CNN models applied to Poha is evaluated and compared with the proposed model. The obtained experimental results show that the proposed model is superior to other models with test accuracy of 99.64 percent for PHCR. The results indicate that our model may be a strong candidate for handwritten character recognition and automated PHCR applications.

12.
Sensors (Basel) ; 19(1)2019 Jan 02.
Artículo en Inglés | MEDLINE | ID: mdl-30609715

RESUMEN

In the era of the Internet of Things and Artificial Intelligence, the Wi-Fi fingerprinting-based indoor positioning system (IPS) has been recognized as the most promising IPS for various applications. Fingerprinting-based algorithms critically rely on a fingerprint database built from machine learning methods. However, currently methods are based on single-feature Received Signal Strength (RSS), which is extremely unstable in performance in terms of precision and robustness. The reason for this is that single feature machines cannot capture the complete channel characteristics and are susceptible to interference. The objective of this paper is to exploit the Time of Arrival (TOA) feature and propose a heterogeneous features fusion model to enhance the precision and robustness of indoor positioning. Several challenges are addressed: (1) machine learning models based on heterogeneous features, (2) the optimization of algorithms for high precision and robustness, and (3) computational complexity. This paper provides several heterogeneous features fusion-based localization models. Their effectiveness and efficiency are thoroughly compared with state-of-the-art methods.

13.
Sensors (Basel) ; 19(1)2019 Jan 04.
Artículo en Inglés | MEDLINE | ID: mdl-30621207

RESUMEN

In order to realize automation of the pollutant emission tests of vehicles, a pedal robot is designed instead of a human-driven vehicle. Sometimes, the actual time-speed curve of the vehicle will deviate from the upper or lower limit of the worldwide light-duty test cycle (WLTC) target curve, which will cause a fault. In this paper, a new fault diagnosis method is proposed and applied to the pedal robot. Since principal component analysis (PCA), t-distributed stochastic neighbor embedding (t-SNE), and Autoencoder cannot extract feature information adequately when they are used alone, three types of feature components extracted by PCA, t-SNE, and Autoencoder are fused to form a nine-dimensional feature set. Then, the feature set is reduced into three-dimensional space via Treelet Transform. Finally, the fault samples are classified by Gaussian process classifier. Compared with the methods using only one algorithm to extract features, the proposed method has the minimum standard deviation, 0.0078, and almost the maximum accuracy, 98.17%. The accuracy of the proposed method is only 0.24% lower than that without Treelet Transform, but the processing time is 6.73% less than that without Treelet Transform. These indicate that the multi-features fusion model and Treelet Transform method is quite effective. Therefore, the proposed method is quite helpful for fault diagnosis of the pedal robot.

14.
Sensors (Basel) ; 19(5)2019 Mar 12.
Artículo en Inglés | MEDLINE | ID: mdl-30871046

RESUMEN

With the revolutionary development of cloud computing and internet of things, the integration and utilization of "big data" resources is a hot topic of the artificial intelligence research. Face recognition technology information has the advantages of being non-replicable, non-stealing, simple and intuitive. Video face tracking in the context of big data has become an important research hotspot in the field of information security. In this paper, a multi-feature fusion adaptive adjustment target tracking window and an adaptive update template particle filter tracking framework algorithm are proposed. Firstly, the skin color and edge features of the face are extracted in the video sequence. The weighted color histogram are extracted which describes the face features. Then we use the integral histogram method to simplify the histogram calculation of the particles. Finally, according to the change of the average distance, the tracking window is adjusted to accurately track the tracking object. At the same time, the algorithm can adaptively update the tracking template which improves the accuracy and accuracy of the tracking. The experimental results show that the proposed method improves the tracking effect and has strong robustness in complex backgrounds such as skin color, illumination changes and face occlusion.

15.
J Med Syst ; 43(12): 329, 2019 Nov 01.
Artículo en Inglés | MEDLINE | ID: mdl-31676931

RESUMEN

Doctor utilizes various kinds of clinical technologies like MRI, endoscopy, CT scan, etc., to identify patient's deformity during the review time. Among set of clinical technologies, wireless capsule endoscopy (WCE) is an advanced procedures used for digestive track malformation. During this complete process, more than 57,000 frames are captured and doctors need to examine a complete video frame by frame which is a tedious task even for an experienced gastrologist. In this article, a novel computerized automated method is proposed for the classification of abdominal infections of gastrointestinal track from WCE images. Three core steps of the suggested system belong to the category of segmentation, deep features extraction and fusion followed by robust features selection. The ulcer abnormalities from WCE videos are initially extracted through a proposed color features based low level and high-level saliency (CFbLHS) estimation method. Later, DenseNet CNN model is utilized and through transfer learning (TL) features are computed prior to feature optimization using Kapur's entropy. A parallel fusion methodology is opted for the selection of maximum feature value (PMFV). For feature selection, Tsallis entropy is calculated later sorted into descending order. Finally, top 50% high ranked features are selected for classification using multilayered feedforward neural network classifier for recognition. Simulation is performed on collected WCE dataset and achieved maximum accuracy of 99.5% in 21.15 s.


Asunto(s)
Endoscopía Capsular/métodos , Hemorragia/diagnóstico , Procesamiento de Imagen Asistido por Computador/métodos , Redes Neurales de la Computación , Gastropatías/diagnóstico , Hemorragia/diagnóstico por imagen , Humanos , Gastropatías/diagnóstico por imagen , Úlcera Gástrica/diagnóstico por imagen
16.
BMC Cancer ; 18(1): 638, 2018 Jun 05.
Artículo en Inglés | MEDLINE | ID: mdl-29871593

RESUMEN

BACKGROUND: Melanoma is the deadliest type of skin cancer with highest mortality rate. However, the annihilation in its early stage implies a high survival rate therefore, it demands early diagnosis. The accustomed diagnosis methods are costly and cumbersome due to the involvement of experienced experts as well as the requirements for the highly equipped environment. The recent advancements in computerized solutions for this diagnosis are highly promising with improved accuracy and efficiency. METHODS: In this article, a method for the identification and classification of the lesion based on probabilistic distribution and best features selection is proposed. The probabilistic distribution such as normal distribution and uniform distribution are implemented for segmentation of lesion in the dermoscopic images. Then multi-level features are extracted and parallel strategy is performed for fusion. A novel entropy-based method with the combination of Bhattacharyya distance and variance are calculated for the selection of best features. Only selected features are classified using multi-class support vector machine, which is selected as a base classifier. RESULTS: The proposed method is validated on three publicly available datasets such as PH2, ISIC (i.e. ISIC MSK-2 and ISIC UDA), and Combined (ISBI 2016 and ISBI 2017), including multi-resolution RGB images and achieved accuracy of 97.5%, 97.75%, and 93.2%, respectively. CONCLUSION: The base classifier performs significantly better on proposed features fusion and selection method as compared to other methods in terms of sensitivity, specificity, and accuracy. Furthermore, the presented method achieved satisfactory segmentation results on selected datasets.


Asunto(s)
Dermoscopía/métodos , Interpretación de Imagen Asistida por Computador/métodos , Melanoma/diagnóstico por imagen , Neoplasias Cutáneas/diagnóstico por imagen , Algoritmos , Humanos , Melanoma/clasificación , Neoplasias Cutáneas/clasificación
17.
Sheng Wu Yi Xue Gong Cheng Xue Za Zhi ; 35(1): 15-24, 2018 02 25.
Artículo en Zh | MEDLINE | ID: mdl-29745595

RESUMEN

To improve the performance of brain-controlled intelligent car based on motor imagery (MI), a method based on neurofeedback (NF) with electroencephalogram (EEG) for controlling intelligent car is proposed. A mental strategy of MI in which the energy column diagram of EEG features related to the mental activity is presented to subjects with visual feedback in real time to train them to quickly master the skills of MI and regulate their EEG activity, and combination of multi-features fusion of MI and multi-classifiers decision were used to control the intelligent car online. The average, maximum and minimum accuracy of identifying instructions achieved by the trained group (trained by the designed feedback system before the experiment) were 85.71%, 90.47% and 76.19%, respectively and the corresponding accuracy achieved by the control group (untrained) were 73.32%, 80.95% and 66.67%, respectively. For the trained group, the average, longest and shortest time consuming were 92 s, 101 s, and 85 s, respectively, while for the control group the corresponding time were 115.7 s, 120 s, and 110 s, respectively. According to the results described above, it is expected that this study may provide a new idea for the follow-up development of brain-controlled intelligent robot by the neurofeedback with EEG related to MI.

18.
Neural Netw ; 173: 106144, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38335792

RESUMEN

The current models for the salient object detection (SOD) have made remarkable progress through multi-scale feature fusion strategies. However, the existing models have large deviations in the detection of different scales, and the target boundaries of the prediction images are still blurred. In this paper, we propose a new model addressing these issues using a transformer backbone to capture multiple feature layers. The model uses multi-scale skip residual connections during encoding to improve the accuracy of the model's predicted object position and edge pixel information. Furthermore, to extract richer multi-scale semantic information, we perform multiple mixed feature operations in the decoding stage. In addition, we add the structure similarity index measure (SSIM) function with coefficients in the loss function to enhance the accurate prediction performance of the boundaries. Experiments demonstrate that our algorithm achieves state-of-the-art results on five public datasets, and improves the performance metrics of the existing SOD tasks. Codes and results are available at: https://github.com/xxwudi508/MSRMNet.


Asunto(s)
Algoritmos , Benchmarking , Semántica
19.
Comput Biol Med ; 169: 107911, 2024 Feb.
Artículo en Inglés | MEDLINE | ID: mdl-38160501

RESUMEN

Extracting expressive molecular features is essential for molecular property prediction. Sequence-based representation is a common representation of molecules, which ignores the structure information of molecules. While molecular graph representation has a weak ability in expressing the 3D structure. In this article, we try to make use of the advantages of different type representations simultaneously for molecular property prediction. Thus, we propose a fusion model named DLF-MFF, which integrates the multi-type molecular features. Specifically, we first extract four different types of features from molecular fingerprints, 2D molecular graph, 3D molecular graph and molecular image. Then, in order to learn molecular features individually, we use four essential deep learning frameworks, which correspond to four distinct molecular representations. The final molecular representation is created by integrating the four feature vectors and feeding them into prediction layer to predict molecular property. We compare DLF-MFF with 7 state-of-the-art methods on 6 benchmark datasets consisting of multiple molecular properties, the experimental results show that DLF-MFF achieves state-of-the-art performance on 6 benchmark datasets. Moreover, DLF-MFF is applied to identify potential anti-SARS-CoV-2 inhibitor from 2500 drugs. We predict probability of each drug being inferred as a 3CL protease inhibitor and also calculate the binding affinity scores between each drug and 3CL protease. The results show that DLF-MFF product better performance in the identification of anti-SARS-CoV-2 inhibitor. This work is expected to offer novel research perspectives for accurate prediction of molecular properties and provide valuable insights into drug repurposing for COVID-19.


Asunto(s)
COVID-19 , Aprendizaje Profundo , Humanos , Antivirales , Benchmarking , Reposicionamiento de Medicamentos , SARS-CoV-2
20.
PeerJ Comput Sci ; 10: e1812, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38259889

RESUMEN

Object detection plays an important role in the field of computer vision. The purpose of object detection is to identify the objects of interest in the image and determine their categories and positions. Object detection has many important applications in various fields. This article addresses the problems of unclear foreground contour in moving object detection and excessive noise points in the global vision, proposing an improved Gaussian mixture model for feature fusion. First, the RGB image was converted into the HSV space, and a mixed Gaussian background model was established. Next, the object area was obtained through background subtraction, residual interference in the foreground was removed using the median filtering method, and morphological processing was performed. Then, an improved Canny algorithm using an automatic threshold from the Otsu method was used to extract the overall object contour. Finally, feature fusion of edge contours and the foreground area was performed to obtain the final object contour. The experimental results show that this method improves the accuracy of the object contour and reduces noise in the object.

SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA