RESUMO
Detecting small objects in images poses significant challenges due to their limited pixel representation and the difficulty in extracting sufficient features, often leading to missed or false detections. To address these challenges and enhance detection accuracy, this paper presents an improved small object detection algorithm, CRL-YOLOv5. The proposed approach integrates the Convolutional Block Attention Module (CBAM) attention mechanism into the C3 module of the backbone network, which enhances the localization accuracy of small objects. Additionally, the Receptive Field Block (RFB) module is introduced to expand the model's receptive field, thereby fully leveraging contextual information. Furthermore, the network architecture is restructured to include an additional detection layer specifically for small objects, allowing for deeper feature extraction from shallow layers. When tested on the VisDrone2019 small object dataset, CRL-YOLOv5 achieved an mAP50 of 39.2%, representing a 5.4% improvement over the original YOLOv5, effectively boosting the detection precision for small objects in images.
RESUMO
Empathy determines our emotional and social lives. Research has recognized the role of the right temporoparietal junction (rTPJ) in social cognition; however, there is less direct causal evidence for its involvement in empathic responses to pain, which is typically attributed to simulation mechanisms. Given the rTPJ's role in processing false beliefs and contextual information during social scenarios, we hypothesized that empathic responses to another person's pain depend on the rTPJ if participants are given information about people's intentions, engaging mentalizing mechanisms alongside simulative ones. Participants viewed videos of an actress freely showing or suppressing pain caused by an electric shock while receiving 6 Hz repetitive transcranial magnetic stimulation (rTMS) over the rTPJ or sham vertex stimulation. Active rTMS had no significant effect on participants' ratings depending on the pain expression, although participants rated the actress's pain as lower during rTPJ perturbation. In contrast, rTMS accelerated response times for providing ratings during pain suppression. We also found that participants perceived the actress's pain as more intense when they knew she would suppress it rather than show it. These results suggest an involvement of the rTPJ in attributing pain to others and provide new insights into people's behavior in judging others' pain when it is concealed.
Assuntos
Empatia , Dor , Lobo Parietal , Estimulação Magnética Transcraniana , Humanos , Empatia/fisiologia , Feminino , Masculino , Estimulação Magnética Transcraniana/métodos , Adulto Jovem , Dor/psicologia , Dor/fisiopatologia , Adulto , Lobo Parietal/fisiologia , Lobo Temporal/fisiologia , Cognição/fisiologiaRESUMO
Vision Transformer have achieved impressive performance in image super-resolution. However, they suffer from low inference speed mainly because of the quadratic complexity of multi-head self-attention (MHSA), which is the key to learning long-range dependencies. On the contrary, most CNN-based methods neglect the important effect of global contextual information, resulting in inaccurate and blurring details. If one can make the best of both Transformers and CNNs, it will achieve a better trade-off between image quality and inference speed. Based on this observation, firstly assume that the main factor affecting the performance in the Transformer-based SR models is the general architecture design, not the specific MHSA component. To verify this, some ablation studies are made by replacing MHSA with large kernel convolutions, alongside other essential module replacements. Surprisingly, the derived models achieve competitive performance. Therefore, a general architecture design GlobalSR is extracted by not specifying the core modules including blocks and domain embeddings of Transformer-based SR models. It also contains three practical guidelines for designing a lightweight SR network utilizing image-level global contextual information to reconstruct SR images. Following the guidelines, the blocks and domain embeddings of GlobalSR are instantiated via Deformable Convolution Attention Block (DCAB) and Fast Fourier Convolution Domain Embedding (FCDE), respectively. The instantiation of general architecture, termed GlobalSR-DF, proposes a DCA to extract the global contextual feature by utilizing Deformable Convolution and a Hadamard product as the attention map at the block level. Meanwhile, the FCDE utilizes the Fast Fourier to transform the input spatial feature into frequency space and then extract image-level global information from it by convolutions. Extensive experiments demonstrate that GlobalSR is the key part in achieving a superior trade-off between SR quality and efficiency. Specifically, our proposed GlobalSR-DF outperforms state-of-the-art CNN-based and ViT-based SISR models regarding accuracy-speed trade-offs with sharp and natural details.
RESUMO
Cortical neurons encode both sensory and contextual information, yet it remains unclear how experiences modulate these cortical representations. Here, we demonstrate that trace eyeblink conditioning (TEC), an aversive associative-learning paradigm linking conditioned (CS) with unconditioned stimuli (US), finely tunes cortical coding at both population and single-neuron levels. Initially, we show that the primary somatosensory cortex (S1) is necessary for TEC acquisition, as evidenced by local muscimol administration. At the population level, TEC enhances activity in a small subset (â¼20%) of CS- or US-responsive primary neurons (rPNs) while diminishing activity in non-rPNs, including locomotion-tuned or unresponsive PNs. Crucially, TEC learning modulates the encoding of sensory versus contextual information in single rPNs: CS-responsive neurons become less responsive, while US-responsive neurons gain responses to CS. Moreover, we find that the cholinergic pathway, via nicotinic receptors, underlies TEC-induced modulations. These findings suggest that experiences dynamically tune cortical representations through cholinergic pathways.
Assuntos
Córtex Somatossensorial , Animais , Córtex Somatossensorial/fisiologia , Masculino , Neurônios/fisiologia , Neurônios/metabolismo , Camundongos , Aprendizagem por Associação/fisiologia , Feminino , Condicionamento Clássico/fisiologia , Condicionamento Palpebral/fisiologia , Camundongos Endogâmicos C57BL , Muscimol/farmacologia , Receptores Nicotínicos/metabolismoRESUMO
Water pollution has become a major concern in recent years, affecting over 2 billion people worldwide, according to UNESCO. This pollution can occur by either naturally, such as algal blooms, or man-made when toxic substances are released into water bodies like lakes, rivers, springs, and oceans. To address this issue and monitor surface-level water pollution in local water bodies, an informative real-time vision-based surveillance system has been developed in conjunction with large language models (LLMs). This system has an integrated camera connected to a Raspberry Pi for processing input frames and is further linked to LLMs for generating contextual information regarding the type, causes, and impact of pollutants on both human health and the environment. This multi-model setup enables local authorities to monitor water pollution and take necessary steps to mitigate it. To train the vision model, seven major types of pollutants found in water bodies like algal bloom, synthetic foams, dead fishes, oil spills, wooden logs, industrial waste run-offs, and trashes were used for achieving accurate detection. ChatGPT API has been integrated with the model to generate contextual information about pollution detected. Thus, the multi-model system can conduct surveillance over water bodies and autonomously alert local authorities to take immediate action, eliminating the need for human intervention. PRACTITIONER POINTS: Combines cameras and LLMs with Raspberry Pi for processing and generating pollutant information. Uses YOLOv5 to detect algal blooms, synthetic foams, dead fish, oil spills, and industrial waste. Supports various modules and environments, including drones and mobile apps for broad monitoring. Educates on environmental healthand alerts authorities about water pollution.
Assuntos
Monitoramento Ambiental , Poluição da Água , Monitoramento Ambiental/métodos , Poluição da Água/análise , Inteligência Artificial , Modelos TeóricosRESUMO
BACKGROUND: Precise glioma segmentation from multi-parametric magnetic resonance (MR) images is essential for brain glioma diagnosis. However, due to the indistinct boundaries between tumor sub-regions and the heterogeneous appearances of gliomas in volumetric MR scans, designing a reliable and automated glioma segmentation method is still challenging. Although existing 3D Transformer-based or convolution-based segmentation networks have obtained promising results via multi-modal feature fusion strategies or contextual learning methods, they widely lack the capability of hierarchical interactions between different modalities and cannot effectively learn comprehensive feature representations related to all glioma sub-regions. PURPOSE: To overcome these problems, in this paper, we propose a 3D hierarchical cross-modality interaction network (HCMINet) using Transformers and convolutions for accurate multi-modal glioma segmentation, which leverages an effective hierarchical cross-modality interaction strategy to sufficiently learn modality-specific and modality-shared knowledge correlated to glioma sub-region segmentation from multi-parametric MR images. METHODS: In the HCMINet, we first design a hierarchical cross-modality interaction Transformer (HCMITrans) encoder to hierarchically encode and fuse heterogeneous multi-modal features by Transformer-based intra-modal embeddings and inter-modal interactions in multiple encoding stages, which effectively captures complex cross-modality correlations while modeling global contexts. Then, we collaborate an HCMITrans encoder with a modality-shared convolutional encoder to construct the dual-encoder architecture in the encoding stage, which can learn the abundant contextual information from global and local perspectives. Finally, in the decoding stage, we present a progressive hybrid context fusion (PHCF) decoder to progressively fuse local and global features extracted by the dual-encoder architecture, which utilizes the local-global context fusion (LGCF) module to efficiently alleviate the contextual discrepancy among the decoding features. RESULTS: Extensive experiments are conducted on two public and competitive glioma benchmark datasets, including the BraTS2020 dataset with 494 patients and the BraTS2021 dataset with 1251 patients. Results show that our proposed method outperforms existing Transformer-based and CNN-based methods using other multi-modal fusion strategies in our experiments. Specifically, the proposed HCMINet achieves state-of-the-art mean DSC values of 85.33% and 91.09% on the BraTS2020 online validation dataset and the BraTS2021 local testing dataset, respectively. CONCLUSIONS: Our proposed method can accurately and automatically segment glioma regions from multi-parametric MR images, which is beneficial for the quantitative analysis of brain gliomas and helpful for reducing the annotation burden of neuroradiologists.
RESUMO
It is now well established that decision making can be susceptible to cognitive bias in a broad range of fields, with forensic science being no exception. Previously published research has revealed a bias blind spot in forensic science where examiners do not recognise bias within their own domain. A survey of 101 forensic anthropology practitioners (n = 52) and students (n = 38) was undertaken to assess their level of awareness of cognitive bias and investigate their attitudes towards cognitive bias within forensic anthropology. The results revealed that the forensic anthropology community (â¼90%) had a high level of awareness of cognitive bias. Overall â¼89% expressed concerns about cognitive bias in the broad discipline of forensic science, their own domain of forensic anthropology, and in the evaluative judgments they made in reconstruction activities, identifying a significant reduction in the bias blind spot. However, more than half of the participants believed that bias can be reduced by sheer force of will, and there was a lack of consensus about implementing blinding procedures or context management. These findings highlight the need to investigate empirically the feasibility of proposed mitigating strategies within the workflow of forensic anthropologists and their capabilities for increasing the transparency in decision making.
Assuntos
Atitude , Antropologia Forense , Humanos , Antropologia Forense/métodos , Inquéritos e Questionários , Masculino , Feminino , Viés , Cognição , Tomada de Decisões , AdultoRESUMO
Gamma Knife radiosurgery (GKRS) is a well-established technique in radiation therapy (RT) for treating small-size brain tumors. It administers highly concentrated doses during each treatment fraction, with even minor dose errors posing a significant risk of causing severe damage to healthy tissues. It underscores the critical need for precise and meticulous precision in GKRS. However, the planning process for GKRS is complex and time-consuming, heavily reliant on the expertise of medical physicists. Incorporating deep learning approaches for GKRS dose prediction can reduce this dependency, improve planning efficiency and homogeneity, streamline clinical workflows, and reduce patient lagging times. Despite this, precise Gamma Knife plan dose distribution prediction using existing models remains a significant challenge. The complexity stems from the intricate nature of dose distributions, subtle contrasts in CT scans, and the interdependence of dosimetric metrics. To overcome these challenges, we have developed a "Cascaded-Deep-Supervised" Convolutional Neural Network (CDS-CNN) that employs a hybrid-weighted optimization scheme. Our innovative method incorporates multi-level deep supervision and a strategic sequential multi-network training approach. It enables the extraction of intra-slice and inter-slice features, leading to more realistic dose predictions with additional contextual information. CDS-CNN was trained and evaluated using data from 105 brain cancer patients who underwent GKRS treatment, with 85 cases used for training and 20 for testing. Quantitative assessments and statistical analyses demonstrated high consistency between the predicted dose distributions and the reference doses from the treatment planning system (TPS). The 3D overall gamma passing rates (GPRs) reached 97.15% ± 1.36% (3 mm/3%, 10% threshold), surpassing the previous best performance by 2.53% using the 3D Dense U-Net model. When evaluated against more stringent criteria (2 mm/3%, 10% threshold, and 1 mm/3%, 10% threshold), the overall GPRs still achieved 96.53% ± 1.08% and 95.03% ± 1.18%. Furthermore, the average target coverage (TC) was 98.33% ± 1.16%, dose selectivity (DS) was 0.57 ± 0.10, gradient index (GI) was 2.69 ± 0.30, and homogeneity index (HI) was 1.79 ± 0.09. Compared to the 3D Dense U-Net, CDS-CNN predictions demonstrated a 3.5% improvement in TC, and CDS-CNN's dose prediction yielded better outcomes than the 3D Dense U-Net across all evaluation criteria. The experimental results demonstrated that the proposed CDS-CNN model outperformed other models in predicting GKRS dose distributions, with predictions closely matching the TPS doses.
RESUMO
Objective. Convolutional neural network (CNN) is developing rapidly in the field of medical image registration, and the proposed U-Net further improves the precision of registration. However, this method may discard certain important information in the process of encoding and decoding steps, consequently leading to a decline in accuracy. To solve this problem, a multi-channel semantic-aware and residual attention mechanism network (MSRA-Net) is proposed in this paper.Approach. Our proposed network achieves efficient information aggregation by cleverly extracting the features of different channels. Firstly, a context-aware module (CAM) is designed to extract valuable contextual information. And the depth-wise separable convolution is employed in the CAM to alleviate the computational burden. Then, a new multi-channel semantic-aware module (MCSAM) is designed for more comprehensive fusion of up-sampling features. Additionally, the residual attention module is introduced in the up-sampling process to extract more semantic information and minimize information loss.Main results. This study utilizes Dice score, average symmetric surface distance and negative Jacobian determinant evaluation metrics to evaluate the influence of registration. The experimental results demonstrate that our proposed MSRA-Net has the highest accuracy compared to several state-of-the-art methods. Moreover, our network has demonstrated the highest Dice score across multiple datasets, thereby indicating that the superior generalization capabilities of our model.Significance. The proposed MSRA-Net offers a novel approach to improve medical image registration accuracy, with implications for various clinical applications. Our implementation is available athttps://github.com/shy922/MSRA-Net.
Assuntos
Imageamento Tridimensional , Redes Neurais de Computação , Semântica , Imageamento Tridimensional/métodos , Humanos , Aprendizado de Máquina não SupervisionadoRESUMO
Unsupervised domain adaptation (UDA) aims to reapply the classifier to be ever-trained on a labeled source domain to a related unlabeled target domain. Recent progress in this line has evolved with the advance of network architectures from convolutional neural networks (CNNs) to transformers or both hybrids. However, this advance has to pay the cost of high computational overheads or complex training processes. In this paper, we propose an efficient alternative hybrid architecture by marrying transformer to contextual convolution (TransConv) to solve UDA tasks. Different from previous transformer based UDA architectures, TransConv has two special aspects: (1) reviving the multilayer perception (MLP) of transformer encoders with Gaussian channel attention fusion for robustness, and (2) mixing contextual features to highly efficient dynamic convolutions for cross-domain interaction. As a result, TransConv enables to calibrate interdomain feature semantics from the global features and the local ones. Experimental results on five benchmarks show that TransConv attains remarkable results with high efficiency as compared to the existing UDA methods.
RESUMO
In sport, coaches often explicitly provide athletes with stable contextual information related to opponent action preferences to enhance anticipation performance. This information can be dependent on, or independent of, dynamic contextual information that only emerges during the sequence of play (e.g. opponent positioning). The interdependency between contextual information sources, and the associated cognitive demands of integrating information sources during anticipation, has not yet been systematically examined. We used a temporal occlusion paradigm to alter the reliability of contextual and kinematic information during the early, mid- and final phases of a two-versus-two soccer anticipation task. A dual-task paradigm was incorporated to investigate the impact of task load on skilled soccer players' ability to integrate information and update their judgements in each phase. Across conditions, participants received no contextual information (control) or stable contextual information (opponent preferences) that was dependent on, or independent of, dynamic contextual information (opponent positioning). As predicted, participants used reliable contextual and kinematic information to enhance anticipation. Further exploratory analysis suggested that increased task load detrimentally affected anticipation accuracy but only when both reliable contextual and kinematic information were available for integration in the final phase. This effect was observed irrespective of whether the stable contextual information was dependent on, or independent of, dynamic contextual information. Findings suggest that updating anticipatory judgements in the final phase of a sequence of play based on the integration of reliable contextual and kinematic information requires cognitive resources.
Assuntos
Atletas , Futebol , Humanos , Reprodutibilidade dos Testes , Fonte de Informação , JulgamentoRESUMO
Crowd counting is an important task that serves as a preprocessing step in many applications. Despite obvious improvement reported by various convolutional-neural-network-based approaches, they only focus on the role of deep feature maps while neglecting the importance of shallow features for crowd counting. In order to surmount this issue, a dilated convolutional-neural-network-based cross-level contextual information extraction network is proposed in this work, which is abbreviated as CL-DCNN. Specifically, a dilated contextual module (DCM) is constructed by importing cross-level connection between different feature maps. It can effectively integrate contextual information while conserving the local details of crowd scenes. Extensive experiments show that the proposed approach outperforms state-of-the-art approaches using five public datasets, i.e., ShanghaiTech part A, ShanghaiTech part B, Mall, UCF_CC_50 and UCF-QNRF, achieving MAE 52.6, 8.1, 1.55, 181.8, and 96.4, respectively.
RESUMO
Accurate segmentation of hepatic vessel is significant for the surgeons to design the preoperative planning of liver surgery. In this paper, a sequence-based context-aware association network (SCAN) is designed for hepatic vessel segmentation, in which three schemes are incorporated to simultaneously extract the 2D features of hepatic vessels and capture the correlations between adjacent CT slices. The two schemes of slice-level attention module and graph association module are designed to bridge feature gaps between the encoder and the decoder in the low- and high-dimensional spaces. The region-edge constrained loss is designed to well optimize the proposed SCAN, which integrates cross-entropy loss, dice loss, and edge-constrained loss. Experimental results indicate that the proposed SCAN is superior to several existing deep learning frameworks, in terms of 0.845 DSC, 0.856 precision, 0.866 sensitivity, and 0.861 F1-score.
Assuntos
Cirurgiões , Humanos , Entropia , Processamento de Imagem Assistida por ComputadorRESUMO
Expert performers in time constrained sports use a range of information sources to facilitate anticipatory and decision-making processes. However, research has often focused on responders such as batters, goalkeepers, defenders, and returners of serve, and failed to capture the complex interaction between opponents, where responders can also manipulate probabilities in their favour. This investigation aimed to explore the interaction between top order batters and fast or medium paced bowlers in cricket and the information they use to inform their anticipatory and decision-making skills in Twenty20 competition. Eleven professional cricketers were interviewed (8 batters and 3 bowlers) using semi-structured questions and scenarios from Twenty20 matches. An inductive and deductive thematic analysis was conducted using the overarching themes of Situation Awareness (SA) and Option Awareness (OA). Within SA, the sub-themes identified related to information sources used by bowlers and batters (i.e., stable contextual information, dynamic contextual information, kinematic information). Within OA, the sub-themes identified highlighted how cricketers use these information sources to understand the options available and the likelihood of success associated with each option (e.g., risk and reward, personal strengths). A sub-theme of 'responder manipulation' was also identified within OA to provide insight into how batters and bowlers interact in a cat-and-mouse like manner to generate options that manipulate one another throughout the competition. A schematic has been developed based on the study findings to illustrate the complex interaction between the anticipation and decision-making processes of professional top order batters and fast or medium paced bowlers in Twenty20 cricket.
Assuntos
Críquete , Esportes , Humanos , Fenômenos Biomecânicos , Probabilidade , LogroRESUMO
In the realm of modern medicine, medical imaging stands as an irreplaceable pillar for accurate diagnostics. The significance of precise segmentation in medical images cannot be overstated, especially considering the variability introduced by different practitioners. With the escalating volume of medical imaging data, the demand for automated and efficient segmentation methods has become imperative. This study introduces an innovative approach to heart image segmentation, embedding a multi-scale feature and attention mechanism within an inverted pyramid framework. Recognizing the intricacies of extracting contextual information from low-resolution medical images, our method adopts an inverted pyramid architecture. Through training with multi-scale images and integrating prediction outcomes, we enhance the network's contextual understanding. Acknowledging the consistent patterns in the relative positions of organs, we introduce an attention module enriched with positional encoding information. This module empowers the network to capture essential positional cues, thereby elevating segmentation accuracy. Our research resides at the intersection of medical imaging and sensor technology, emphasizing the foundational role of sensors in medical image analysis. The integration of sensor-generated data showcases the symbiotic relationship between sensor technology and advanced machine learning techniques. Evaluation on two heart datasets substantiates the superior performance of our approach. Metrics such as the Dice coefficient, Jaccard coefficient, recall, and F-measure demonstrate the method's efficacy compared to state-of-the-art techniques. In conclusion, our proposed heart image segmentation method addresses the challenges posed by diverse medical images, offering a promising solution for efficiently processing 2D/3D sensor data in contemporary medical imaging.
Assuntos
Benchmarking , Sinais (Psicologia) , Coração/diagnóstico por imagem , Aprendizado de Máquina , Tecnologia , Processamento de Imagem Assistida por ComputadorRESUMO
The counting of pineapple buds relies on target recognition in estimating pineapple yield using unmanned aerial vehicle (UAV) photography. This research proposes the SFHG-YOLO method, with YOLOv5s as the baseline, to address the practical needs of identifying small objects (pineapple buds) in UAV vision and the drawbacks of existing algorithms in terms of real-time performance and accuracy. Field pineapple buds are small objects that may be detected in high density using a lightweight network model. This model enhances spatial attention and adaptive context information fusion to increase detection accuracy and resilience. To construct the lightweight network model, the first step involves utilizing the coordinate attention module and MobileNetV3. Additionally, to fully leverage feature information across various levels and enhance perception skills for tiny objects, we developed both an enhanced spatial attention module and an adaptive context information fusion module. Experiments were conducted to validate the suggested algorithm's performance in detecting small objects. The SFHG-YOLO model exhibited significant gains in assessment measures, achieving mAP@0.5 and mAP@0.5:0.95 improvements of 7.4% and 31%, respectively, when compared to the baseline model YOLOv5s. Considering the model size and computational cost, the findings underscore the superior performance of the suggested technique in detecting high-density small items. This program offers a reliable detection approach for estimating pineapple yield by accurately identifying minute items.
RESUMO
Driving risk prediction is crucial for safety and risk mitigation. While traditional methods rely on demographic information for insurance pricing, they may not fully capture actual driving behavior. To address this, telematics data has gained popularity. This study focuses on using telematics data and contextual information (e.g., road type, daylight) to represent a driver's style through tensor representations. Drivers with similar behaviors are identified by clustering their representations, forming risk cohorts. Past at-fault traffic accidents and citations serve as partial risk labels. The relative magnitude of average records (per driver) for each cohort indicates their risk label, such as low or high risk, which can be transferred to drivers in a cohort. A classifier is then constructed using augmented risk labels and driving style representations to predict driving risk for new drivers. Real-world data from major US cities validates the effectiveness of this framework. The approach is practical for large-scale scenarios as the data can be obtained at scale. Its focus on driver-based risk prediction makes it applicable to industries like auto-insurance. Beyond personalized premiums, the framework empowers drivers to assess their driving behavior in various contexts, facilitating skill improvement over time.
Assuntos
Acidentes de Trânsito , Conscientização , Humanos , Acidentes de Trânsito/prevenção & controle , Cidades , Análise por Conglomerados , IndústriasRESUMO
In many industrial sectors, workers are exposed to manufactured or unintentionally emitted airborne nanoparticles (NPs). To develop prevention and enhance knowledge surrounding exposure, it has become crucial to achieve a consensus on how to assess exposure to airborne NPs by inhalation in the workplace. Here, we review the literature presenting recommendations on assessing occupational exposure to NPs. The 23 distinct strategies retained were analyzed in terms of the following points: target NPs, objectives, steps, "measurement strategy" (instruments, physicochemical analysis, and data processing), "contextual information" presented, and "work activity" analysis. The robustness (consistency of information) and practical aspects (detailed methodology) of each strategy were estimated. The objectives and methodological steps varied, as did the measurement techniques. Strategies were essentially based on NPs measurement, but improvements could be made to better account for "contextual information" and "work activity". Based on this review, recommendations for an operational strategy were formulated, integrating the work activity with the measurement to provide a more complete assessment of situations leading to airborne NP exposure. These recommendations can be used with the objective of producing homogeneous exposure data for epidemiological purposes and to help improve prevention strategies.
RESUMO
Long non-coding RNAs (lncRNAs) play a crucial role in numbers of biological processes and have received wide attention during the past years. Since the rapid development of high-throughput transcriptome sequencing technologies (RNA-seq) lead to a large amount of RNA data, it is urgent to develop a fast and accurate coding potential predictor. Many computational methods have been proposed to address this issue, they usually exploit information on open reading frame (ORF), protein sequence, k-mer, evolutionary signatures, or homology. Despite the effectiveness of these approaches, there is still much room to improve. Indeed, none of these methods exploit the contextual information of RNA sequence, for example, k-mer features that counts the occurrence frequencies of continuous nucleotides (k-mer) in the whole RNA sequence cannot reflect local contextual information of each k-mer. In view of this shortcoming, here, we present a novel alignment-free method, CPPVec, which exploits the contextual information of RNA sequence for coding potential prediction for the first time, it can be easily implemented by distributed representation (e.g., doc2vec) of protein sequence translated from the longest ORF. The experimental findings demonstrate that CPPVec is an accurate coding potential predictor and significantly outperforms existing state-of-the-art methods.
Assuntos
RNA Longo não Codificante , Sequência de Aminoácidos , RNA Longo não Codificante/genética , Sequência de Bases , Análise de Sequência de RNARESUMO
The process of image segmentation is partitioning an image into its constituent parts and is a significant approach for extracting interesting features from images. Over a couple of decades, many efficient image segmentation approaches have been formulated for various applications. Still, it is a challenging and complex issue, especially for color image segmentation. To moderate this difficulty, a novel multilevel thresholding approach is proposed in this paper based on the electromagnetism optimization (EMO) technique with an energy curve, named multilevel thresholding based on EMO and energy curve (MTEMOE). To compute the optimized threshold values, Otsu's variance and Kapur's entropy are deployed as fitness functions; both values should be maximized to locate optimal threshold values. In both Kapur's and Otsu's methods, the pixels of an image are classified into different classes based on the threshold level selected on the histogram. Optimal threshold levels give higher efficiency of segmentation; the EMO technique is used to find optimal thresholds in this research. The methods based on an image's histograms do not possess the spatial contextual information for finding the optimal threshold levels. To abolish this deficiency an energy curve is used instead of the histogram and this curve can establish the spatial relationship of pixels with their neighbor pixels. To study the experimental results of the proposed scheme, several color benchmark images are considered at various threshold levels and compared with other meta-heuristic algorithms: multi-verse optimization, whale optimization algorithm, and so on. The investigational results are illustrated in terms of mean square error, peak signal-to-noise ratio, the mean value of fitness reach, feature similarity, structural similarity, variation of information, and probability rand index. The results reveal that the proposed MTEMOE approach overtops other state-of-the-art algorithms to solve engineering problems in various fields.