Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 128.984
Filtrar
1.
Protein Sci ; 33(6): e4985, 2024 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-38717278

RESUMEN

Inteins are proteins that excise themselves out of host proteins and ligate the flanking polypeptides in an auto-catalytic process called protein splicing. In nature, inteins are either contiguous or split. In the case of split inteins, the two fragments must first form a complex for the splicing to occur. Contiguous inteins have previously been artificially split in two fragments because split inteins allow for distinct applications than contiguous ones. Even naturally split inteins have been split at unnatural split sites to obtain fragments with reduced affinity for one another, which are useful to create conditional inteins or to study protein-protein interactions. So far, split sites in inteins have been heuristically identified. We developed Int&in, a web server freely available for academic research (https://intein.biologie.uni-freiburg.de) that runs a machine learning model using logistic regression to predict active and inactive split sites in inteins with high accuracy. The model was trained on a dataset of 126 split sites generated using the gp41-1, Npu DnaE and CL inteins and validated using 97 split sites extracted from the literature. Despite the limited data size, the model, which uses various protein structural features, as well as sequence conservation information, achieves an accuracy of 0.79 and 0.78 for the training and testing sets, respectively. We envision Int&in will facilitate the engineering of novel split inteins for applications in synthetic and cell biology.


Asunto(s)
Inteínas , Internet , Aprendizaje Automático , Empalme de Proteína , Programas Informáticos , Dominio Catalítico
2.
PLoS Comput Biol ; 20(5): e1012024, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38717988

RESUMEN

The activation levels of biologically significant gene sets are emerging tumor molecular markers and play an irreplaceable role in the tumor research field; however, web-based tools for prognostic analyses using it as a tumor molecular marker remain scarce. We developed a web-based tool PESSA for survival analysis using gene set activation levels. All data analyses were implemented via R. Activation levels of The Molecular Signatures Database (MSigDB) gene sets were assessed using the single sample gene set enrichment analysis (ssGSEA) method based on data from the Gene Expression Omnibus (GEO), The Cancer Genome Atlas (TCGA), The European Genome-phenome Archive (EGA) and supplementary tables of articles. PESSA was used to perform median and optimal cut-off dichotomous grouping of ssGSEA scores for each dataset, relying on the survival and survminer packages for survival analysis and visualisation. PESSA is an open-access web tool for visualizing the results of tumor prognostic analyses using gene set activation levels. A total of 238 datasets from the GEO, TCGA, EGA, and supplementary tables of articles; covering 51 cancer types and 13 survival outcome types; and 13,434 tumor-related gene sets are obtained from MSigDB for pre-grouping. Users can obtain the results, including Kaplan-Meier analyses based on the median and optimal cut-off values and accompanying visualization plots and the Cox regression analyses of dichotomous and continuous variables, by selecting the gene set markers of interest. PESSA (https://smuonco.shinyapps.io/PESSA/ OR http://robinl-lab.com/PESSA) is a large-scale web-based tumor survival analysis tool covering a large amount of data that creatively uses predefined gene set activation levels as molecular markers of tumors.


Asunto(s)
Biomarcadores de Tumor , Biología Computacional , Bases de Datos Genéticas , Internet , Neoplasias , Programas Informáticos , Humanos , Neoplasias/genética , Neoplasias/mortalidad , Análisis de Supervivencia , Biomarcadores de Tumor/genética , Biomarcadores de Tumor/metabolismo , Biología Computacional/métodos , Pronóstico , Perfilación de la Expresión Génica/métodos , Regulación Neoplásica de la Expresión Génica/genética
3.
PLoS One ; 19(5): e0298192, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38717996

RESUMEN

Area cartograms are map-based data visualizations in which the area of each map region is proportional to the data value it represents. Long utilized in print media, area cartograms have also become increasingly popular online, often accompanying news articles and blog posts. Despite their popularity, there is a dearth of cartogram generation tools accessible to non-technical users unfamiliar with Geographic Information Systems software. Few tools support the generation of contiguous cartograms (i.e., area cartograms that faithfully represent the spatial adjacency of neighboring regions). We thus reviewed existing contiguous cartogram software and compared two web-based cartogram tools: fBlog and go-cart.io. We experimentally evaluated their usability through a user study comprising cartogram generation and analysis tasks. The System Usability Scale was adopted to quantify how participants perceived the usability of both tools. We also collected written feedback from participants to determine the main challenges faced while using the software. Participants generally rated go-cart.io as being more usable than fBlog. Compared to fBlog, go-cart.io offers a greater variety of built-in maps and allows importing data values by file upload. Still, our results suggest that even go-cart.io suffers from poor usability because the graphical user interface is complex and data can only be imported as a comma-separated-values file. We also propose changes to go-cart.io and make general recommendations for web-based cartogram tools to address these concerns.


Asunto(s)
Internet , Programas Informáticos , Humanos , Femenino , Masculino , Adulto , Sistemas de Información Geográfica , Interfaz Usuario-Computador , Adulto Joven
4.
Opt Lett ; 49(10): 2621-2624, 2024 May 15.
Artículo en Inglés | MEDLINE | ID: mdl-38748120

RESUMEN

Fluorescence fluctuation super-resolution microscopy (FF-SRM) has emerged as a promising method for the fast, low-cost, and uncomplicated imaging of biological specimens beyond the diffraction limit. Among FF-SRM techniques, super-resolution radial fluctuation (SRRF) microscopy is a popular technique but is prone to artifacts, resulting in low fidelity, especially under conditions of high-density fluorophores. In this Letter, we developed a novel, to the best of our knowledge, combinatory computational super-resolution microscopy method, namely VeSRRF, that demonstrated superior performance in SRRF microscopy. VeSRRF combined intensity and gradient variance reweighted radial fluctuations (VRRF) and enhanced-SRRF (eSRRF) algorithms, leveraging the enhanced resolution achieved through intensity and gradient variance analysis in VRRF and the improved fidelity obtained from the radial gradient convergence transform in eSRRF. Our method was validated using microtubules in mammalian cells as a standard biological model system. Our results demonstrated that VeSRRF consistently achieved the highest resolution and exceptional fidelity compared to those obtained from other algorithms in both single-molecule localization microscopy (SMLM) and FF-SRM. Moreover, we developed the VeSRRF software package that is freely available on the open-source ImageJ/Fiji software platform to facilitate the use of VeSRRF in the broader community of biomedical researchers. VeSRRF is an exemplary method in which complementary microscopy techniques are integrated holistically, creating superior imaging performance and capabilities.


Asunto(s)
Algoritmos , Microscopía Fluorescente , Microscopía Fluorescente/métodos , Microtúbulos , Procesamiento de Imagen Asistido por Computador/métodos , Animales , Programas Informáticos
5.
Molecules ; 29(9)2024 May 01.
Artículo en Inglés | MEDLINE | ID: mdl-38731587

RESUMEN

We aimed to obtain the optimal formula for human milk fat substitute (HMFS) through a combination of software and an evaluation model and further verify its practicability through an animal experiment. The results showed that a total of 33 fatty acid (FA) and 63 triglyceride (TAG) molecular species were detected in vegetable oils. Palmitic acid, oleic acid, linoleic acid, 18:1/16:0/18:1, 18:2/16:0/18:2, 18:1/18:1/18:1 and 18:1/18:2/18:1, were the main molecular species among the FAs and TAGs in the vegetable oils. Based on the HMFS evaluation model, the optimal mixed vegetable oil formula was blended with 21.3% palm oil, 2.8% linseed oil, 2.6% soybean oil, 29.9% rapeseed oil and 43.4% maize oil, with the highest score of 83.146. Moreover, there was no difference in the weight, blood routine indices or calcium and magnesium concentrations in the feces of the mice between the homemade mixed vegetable oil (HMVO) group and the commercial mixed vegetable oil (CMVO) group, while nervonic acid (C24:1) and octanoic acid (C8:0) were absorbed easily in the HMVO group. Therefore, these results demonstrate that the mixing of the different vegetable oils was feasible via a combination of computer software and an evaluation model and provided a new way to produce HMFS.


Asunto(s)
Sustitutos de Grasa , Ácidos Grasos , Leche Humana , Aceites de Plantas , Programas Informáticos , Triglicéridos , Humanos , Animales , Aceites de Plantas/química , Ácidos Grasos/química , Leche Humana/química , Ratones , Triglicéridos/química , Sustitutos de Grasa/química , Aceite de Palma/química , Aceite de Soja/química , Aceite de Linaza/química , Aceite de Brassica napus/química , Aceite de Maíz/química , Caprilatos/química , Ácido Palmítico/química , Ácido Oléico/química
6.
Nat Commun ; 15(1): 3992, 2024 May 11.
Artículo en Inglés | MEDLINE | ID: mdl-38734767

RESUMEN

Visual proteomics attempts to build atlases of the molecular content of cells but the automated annotation of cryo electron tomograms remains challenging. Template matching (TM) and methods based on machine learning detect structural signatures of macromolecules. However, their applicability remains limited in terms of both the abundance and size of the molecular targets. Here we show that the performance of TM is greatly improved by using template-specific search parameter optimization and by including higher-resolution information. We establish a TM pipeline with systematically tuned parameters for the automated, objective and comprehensive identification of structures with confidence 10 to 100-fold above the noise level. We demonstrate high-fidelity and high-confidence localizations of nuclear pore complexes, vaults, ribosomes, proteasomes, fatty acid synthases, lipid membranes and microtubules, and individual subunits inside crowded eukaryotic cells. We provide software tools for the generic implementation of our method that is broadly applicable towards realizing visual proteomics.


Asunto(s)
Microscopía por Crioelectrón , Tomografía con Microscopio Electrónico , Complejo de la Endopetidasa Proteasomal , Proteómica , Ribosomas , Programas Informáticos , Tomografía con Microscopio Electrónico/métodos , Microscopía por Crioelectrón/métodos , Ribosomas/ultraestructura , Ribosomas/metabolismo , Complejo de la Endopetidasa Proteasomal/ultraestructura , Complejo de la Endopetidasa Proteasomal/metabolismo , Complejo de la Endopetidasa Proteasomal/química , Humanos , Proteómica/métodos , Poro Nuclear/ultraestructura , Poro Nuclear/metabolismo , Microtúbulos/ultraestructura , Microtúbulos/metabolismo , Ácido Graso Sintasas/metabolismo , Aprendizaje Automático , Imagenología Tridimensional/métodos , Algoritmos , Procesamiento de Imagen Asistido por Computador/métodos
7.
Curr Protoc ; 4(5): e1046, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38717471

RESUMEN

Whole-genome sequencing is widely used to investigate population genomic variation in organisms of interest. Assorted tools have been independently developed to call variants from short-read sequencing data aligned to a reference genome, including single nucleotide polymorphisms (SNPs) and structural variations (SVs). We developed SNP-SVant, an integrated, flexible, and computationally efficient bioinformatic workflow that predicts high-confidence SNPs and SVs in organisms without benchmarked variants, which are traditionally used for distinguishing sequencing errors from real variants. In the absence of these benchmarked datasets, we leverage multiple rounds of statistical recalibration to increase the precision of variant prediction. The SNP-SVant workflow is flexible, with user options to tradeoff accuracy for sensitivity. The workflow predicts SNPs and small insertions and deletions using the Genome Analysis ToolKit (GATK) and predicts SVs using the Genome Rearrangement IDentification Software Suite (GRIDSS), and it culminates in variant annotation using custom scripts. A key utility of SNP-SVant is its scalability. Variant calling is a computationally expensive procedure, and thus, SNP-SVant uses a workflow management system with intermediary checkpoint steps to ensure efficient use of resources by minimizing redundant computations and omitting steps where dependent files are available. SNP-SVant also provides metrics to assess the quality of called variants and converts between VCF and aligned FASTA format outputs to ensure compatibility with downstream tools to calculate selection statistics, which are commonplace in population genomics studies. By accounting for both small and large structural variants, users of this workflow can obtain a wide-ranging view of genomic alterations in an organism of interest. Overall, this workflow advances our capabilities in assessing the functional consequences of different types of genomic alterations, ultimately improving our ability to associate genotypes with phenotypes. © 2024 The Authors. Current Protocols published by Wiley Periodicals LLC. Basic Protocol: Predicting single nucleotide polymorphisms and structural variations Support Protocol 1: Downloading publicly available sequencing data Support Protocol 2: Visualizing variant loci using Integrated Genome Viewer Support Protocol 3: Converting between VCF and aligned FASTA formats.


Asunto(s)
Polimorfismo de Nucleótido Simple , Programas Informáticos , Flujo de Trabajo , Polimorfismo de Nucleótido Simple/genética , Biología Computacional/métodos , Genómica/métodos , Anotación de Secuencia Molecular/métodos , Secuenciación Completa del Genoma/métodos
8.
PLoS One ; 19(5): e0301720, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38739583

RESUMEN

A key benefit of the Open Computing Language (OpenCL) software framework is its capability to operate across diverse architectures. Field programmable gate arrays (FPGAs) are a high-speed computing architecture used for computation acceleration. This study investigates the impact of memory access time on overall performance in general FPGA computing environments through the creation of eight benchmarks within the OpenCL framework. The developed benchmarks capture a range of memory access behaviors, and they play a crucial role in assessing the performance of spinning and sleeping on FPGA-based architectures. The results obtained guide the formulation of new implementations and contribute to defining an abstraction of FPGAs. This abstraction is then utilized to create tailored implementations of primitives that are well-suited for this platform. While other research endeavors concentrate on creating benchmarks with the Compute Unified Device Architecture (CUDA) to scrutinize the memory systems across diverse GPU architectures and propose recommendations for future generations of GPU computation platforms, this study delves into the memory system analysis for the broader FPGA computing platform. It achieves this by employing the highly abstracted OpenCL framework, exploring various data workload characteristics, and experimentally delineating the appropriate implementation of primitives that can seamlessly integrate into a design tailored for the FPGA computing platform. Additionally, the results underscore the efficacy of employing a task-parallel model to mitigate the need for high-cost synchronization mechanisms in designs constructed on general FPGA computing platforms.


Asunto(s)
Benchmarking , Programas Informáticos , Humanos , Lenguajes de Programación
9.
PLoS One ; 19(5): e0291183, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38713711

RESUMEN

BACKGROUND: Mendelian randomisation (MR) is the use of genetic variants as instrumental variables. Mode-based estimators (MBE) are one of the most popular types of estimators used in univariable-MR studies and is often used as a sensitivity analysis for pleiotropy. However, because there are no plurality valid regression estimators, modal estimators for multivariable-MR have been under-explored. METHODS: We use the residual framework for multivariable-MR to introduce two multivariable modal estimators: multivariable-MBE, which uses IVW to create residuals fed into a traditional plurality valid estimator, and an estimator which instead has the residuals fed into the contamination mixture method (CM), multivariable-CM. We then use Monte-Carlo simulations to explore the performance of these estimators when compared to existing ones and re-analyse the data used by Grant and Burgess (2021) looking at the causal effect of intelligence, education, and household income on Alzheimer's disease as an applied example. RESULTS: In our simulation, we found that multivariable-MBE was generally too variable to be much use. Multivariable-CM produced more precise estimates on the other hand. Multivariable-CM performed better than MR-Egger in almost all settings, and Weighted Median under balanced pleiotropy. However, it underperformed Weighted Median when there was a moderate amount of directional pleiotropy. Our re-analysis supported the conclusion of Grant and Burgess (2021), that intelligence had a protective effect on Alzheimer's disease, while education, and household income do not have a causal effect. CONCLUSIONS: Here we introduced two, non-regression-based, plurality valid estimators for multivariable MR. Of these, "multivariable-CM" which uses IVW to create residuals fed into a contamination-mixture model, performed the best. This estimator uses a plurality of variants valid assumption, and appears to provide precise and unbiased estimates in the presence of balanced pleiotropy and small amounts of directional pleiotropy.


Asunto(s)
Análisis de la Aleatorización Mendeliana , Análisis de la Aleatorización Mendeliana/métodos , Humanos , Enfermedad de Alzheimer/genética , Método de Montecarlo , Análisis Multivariante , Simulación por Computador , Variación Genética , Programas Informáticos
10.
Nature ; 629(8012): 509-510, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38719965
11.
Anal Chem ; 96(19): 7634-7642, 2024 May 14.
Artículo en Inglés | MEDLINE | ID: mdl-38691624

RESUMEN

Chemical derivatization is a widely employed strategy in metabolomics to enhance metabolite coverage by improving chromatographic behavior and increasing the ionization rates in mass spectroscopy (MS). However, derivatization might complicate MS data, posing challenges for data mining due to the lack of a corresponding benchmark database. To address this issue, we developed a triple-dimensional combinatorial derivatization strategy for nontargeted metabolomics. This strategy utilizes three structurally similar derivatization reagents and is supported by MS-TDF software for accelerated data processing. Notably, simultaneous derivatization of specific metabolite functional groups in biological samples produced compounds with stable but distinct chromatographic retention times and mass numbers, facilitating discrimination by MS-TDF, an in-house MS data processing software. In this study, carbonyl analogues in human plasma were derivatized using a combination of three hydrazide-based derivatization reagents: 2-hydrazinopyridine, 2-hydrazino-5-methylpyridine, and 2-hydrazino-5-cyanopyridine (6-hydrazinonicotinonitrile). This approach was applied to identify potential carbonyl biomarkers in lung cancer. Analysis and validation of human plasma samples demonstrated that our strategy improved the recognition accuracy of metabolites and reduced the risk of false positives, providing a useful method for nontargeted metabolomics studies. The MATLAB code for MS-TDF is available on GitHub at https://github.com/CaixiaYuan/MS-TDF.


Asunto(s)
Metabolómica , Programas Informáticos , Humanos , Metabolómica/métodos , Neoplasias Pulmonares/metabolismo , Piridinas/química
12.
Nat Commun ; 15(1): 3675, 2024 May 01.
Artículo en Inglés | MEDLINE | ID: mdl-38693118

RESUMEN

The wide applications of liquid chromatography - mass spectrometry (LC-MS) in untargeted metabolomics demand an easy-to-use, comprehensive computational workflow to support efficient and reproducible data analysis. However, current tools were primarily developed to perform specific tasks in LC-MS based metabolomics data analysis. Here we introduce MetaboAnalystR 4.0 as a streamlined pipeline covering raw spectra processing, compound identification, statistical analysis, and functional interpretation. The key features of MetaboAnalystR 4.0 includes an auto-optimized feature detection and quantification algorithm for LC-MS1 spectra processing, efficient MS2 spectra deconvolution and compound identification for data-dependent or data-independent acquisition, and more accurate functional interpretation through integrated spectral annotation. Comprehensive validation studies using LC-MS1 and MS2 spectra obtained from standards mixtures, dilution series and clinical metabolomics samples have shown its excellent performance across a wide range of common tasks such as peak picking, spectral deconvolution, and compound identification with good computing efficiency. Together with its existing statistical analysis utilities, MetaboAnalystR 4.0 represents a significant step toward a unified, end-to-end workflow for LC-MS based global metabolomics in the open-source R environment.


Asunto(s)
Espectrometría de Masas , Metabolómica , Flujo de Trabajo , Algoritmos , Cromatografía Liquida/métodos , Cromatografía Líquida con Espectrometría de Masas , Espectrometría de Masas/métodos , Metabolómica/métodos , Programas Informáticos
13.
Eur Radiol Exp ; 8(1): 63, 2024 May 20.
Artículo en Inglés | MEDLINE | ID: mdl-38764066

RESUMEN

BACKGROUND: Emphysema influences the appearance of lung tissue in computed tomography (CT). We evaluated whether this affects lung nodule detection by artificial intelligence (AI) and human readers (HR). METHODS: Individuals were selected from the "Lifelines" cohort who had undergone low-dose chest CT. Nodules in individuals without emphysema were matched to similar-sized nodules in individuals with at least moderate emphysema. AI results for nodular findings of 30-100 mm3 and 101-300 mm3 were compared to those of HR; two expert radiologists blindly reviewed discrepancies. Sensitivity and false positives (FPs)/scan were compared for emphysema and non-emphysema groups. RESULTS: Thirty-nine participants with and 82 without emphysema were included (n = 121, aged 61 ± 8 years (mean ± standard deviation), 58/121 males (47.9%)). AI and HR detected 196 and 206 nodular findings, respectively, yielding 109 concordant nodules and 184 discrepancies, including 118 true nodules. For AI, sensitivity was 0.68 (95% confidence interval 0.57-0.77) in emphysema versus 0.71 (0.62-0.78) in non-emphysema, with FPs/scan 0.51 and 0.22, respectively (p = 0.028). For HR, sensitivity was 0.76 (0.65-0.84) and 0.80 (0.72-0.86), with FPs/scan of 0.15 and 0.27 (p = 0.230). Overall sensitivity was slightly higher for HR than for AI, but this difference disappeared after the exclusion of benign lymph nodes. FPs/scan were higher for AI in emphysema than in non-emphysema (p = 0.028), while FPs/scan for HR were higher than AI for 30-100 mm3 nodules in non-emphysema (p = 0.009). CONCLUSIONS: AI resulted in more FPs/scan in emphysema compared to non-emphysema, a difference not observed for HR. RELEVANCE STATEMENT: In the creation of a benchmark dataset to validate AI software for lung nodule detection, the inclusion of emphysema cases is important due to the additional number of FPs. KEY POINTS: • The sensitivity of nodule detection by AI was similar in emphysema and non-emphysema. • AI had more FPs/scan in emphysema compared to non-emphysema. • Sensitivity and FPs/scan by the human reader were comparable for emphysema and non-emphysema. • Emphysema and non-emphysema representation in benchmark dataset is important for validating AI.


Asunto(s)
Inteligencia Artificial , Enfisema Pulmonar , Tomografía Computarizada por Rayos X , Humanos , Masculino , Persona de Mediana Edad , Femenino , Tomografía Computarizada por Rayos X/métodos , Enfisema Pulmonar/diagnóstico por imagen , Programas Informáticos , Sensibilidad y Especificidad , Neoplasias Pulmonares/diagnóstico por imagen , Anciano , Dosis de Radiación , Nódulo Pulmonar Solitario/diagnóstico por imagen , Interpretación de Imagen Radiográfica Asistida por Computador/métodos
14.
PLoS One ; 19(5): e0303689, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38768169

RESUMEN

An observational methodology system has been designed which allows the observation and analysis of the technical-tactical behaviour and interaction of judokas during competition. The observation instrument (JUTACTIC) is composed of 8 fixed criteria that provide information related to the competition and the competitors and 13 variable criteria that, throughout the intrasessional monitoring of each combat, allow the behaviour displayed by both judokas and their interaction to be recorded. From an observational sample consisting of matches from the Rio 2016 Olympic champions and the corresponding samples made using the LINCE PLUS software, evidence of validity, reliability, generalizability and applicability of the observation system is provided. The content validity of the observation instrument has been endorsed by a panel of experts (n = 11). Intra and inter-observer reliability has been guaranteed from the results obtained in the Fleiss Kappa and the Krippendorff Alpha. The generalizability analysis with the design structure [Category] [Participants] / [Matches] has confirmed that around seven matches are needed to accurately analyse the behaviour of the competitor under study. The practical application possibilities of the observation instrument has been shown with an example of the results obtained and the regular behaviour structures detected (T-patterns) using the THEME software.


Asunto(s)
Artes Marciales , Humanos , Brasil , Rendimiento Atlético/fisiología , Reproducibilidad de los Resultados , Atletas , Conducta Competitiva , Programas Informáticos , Masculino , Femenino
16.
Brief Bioinform ; 25(3)2024 Mar 27.
Artículo en Inglés | MEDLINE | ID: mdl-38770717

RESUMEN

Drug therapy is vital in cancer treatment. Accurate analysis of drug sensitivity for specific cancers can guide healthcare professionals in prescribing drugs, leading to improved patient survival and quality of life. However, there is a lack of web-based tools that offer comprehensive visualization and analysis of pancancer drug sensitivity. We gathered cancer drug sensitivity data from publicly available databases (GEO, TCGA and GDSC) and developed a web tool called Comprehensive Pancancer Analysis of Drug Sensitivity (CPADS) using Shiny. CPADS currently includes transcriptomic data from over 29 000 samples, encompassing 44 types of cancer, 288 drugs and more than 9000 gene perturbations. It allows easy execution of various analyses related to cancer drug sensitivity. With its large sample size and diverse drug range, CPADS offers a range of analysis methods, such as differential gene expression, gene correlation, pathway analysis, drug analysis and gene perturbation analysis. Additionally, it provides several visualization approaches. CPADS significantly aids physicians and researchers in exploring primary and secondary drug resistance at both gene and pathway levels. The integration of drug resistance and gene perturbation data also presents novel perspectives for identifying pivotal genes influencing drug resistance. Access CPADS at https://smuonco.shinyapps.io/CPADS/ or https://robinl-lab.com/CPADS.


Asunto(s)
Resistencia a Antineoplásicos , Internet , Neoplasias , Programas Informáticos , Humanos , Neoplasias/tratamiento farmacológico , Neoplasias/genética , Resistencia a Antineoplásicos/genética , Antineoplásicos/farmacología , Antineoplásicos/uso terapéutico , Biología Computacional/métodos , Bases de Datos Genéticas , Transcriptoma , Perfilación de la Expresión Génica/métodos
17.
Brief Bioinform ; 25(3)2024 Mar 27.
Artículo en Inglés | MEDLINE | ID: mdl-38770716

RESUMEN

Temporal RNA-sequencing (RNA-seq) studies of bulk samples provide an opportunity for improved understanding of gene regulation during dynamic phenomena such as development, tumor progression or response to an incremental dose of a pharmacotherapeutic. Moreover, single-cell RNA-seq (scRNA-seq) data implicitly exhibit temporal characteristics because gene expression values recapitulate dynamic processes such as cellular transitions. Unfortunately, temporal RNA-seq data continue to be analyzed by methods that ignore this ordinal structure and yield results that are often difficult to interpret. Here, we present Error Modelled Gene Expression Analysis (EMOGEA), a framework for analyzing RNA-seq data that incorporates measurement uncertainty, while introducing a special formulation for those acquired to monitor dynamic phenomena. This method is specifically suited for RNA-seq studies in which low-count transcripts with small-fold changes lead to significant biological effects. Such transcripts include genes involved in signaling and non-coding RNAs that inherently exhibit low levels of expression. Using simulation studies, we show that this framework down-weights samples that exhibit extreme responses such as batch effects allowing them to be modeled with the rest of the samples and maintain the degrees of freedom originally envisioned for a study. Using temporal experimental data, we demonstrate the framework by extracting a cascade of gene expression waves from a well-designed RNA-seq study of zebrafish embryogenesis and an scRNA-seq study of mouse pre-implantation and provide unique biological insights into the regulation of genes in each wave. For non-ordinal measurements, we show that EMOGEA has a much higher rate of true positive calls and a vanishingly small rate of false negative discoveries compared to common approaches. Finally, we provide two packages in Python and R that are self-contained and easy to use, including test data.


Asunto(s)
RNA-Seq , Pez Cebra , Animales , Pez Cebra/genética , RNA-Seq/métodos , Perfilación de la Expresión Génica/métodos , Análisis de la Célula Individual/métodos , Ratones , Análisis de Secuencia de ARN/métodos , Programas Informáticos
18.
Brief Bioinform ; 25(3)2024 Mar 27.
Artículo en Inglés | MEDLINE | ID: mdl-38770718

RESUMEN

Polygenetic Risk Scores are used to evaluate an individual's vulnerability to developing specific diseases or conditions based on their genetic composition, by taking into account numerous genetic variations. This article provides an overview of the concept of Polygenic Risk Scores (PRS). We elucidate the historical advancements of PRS, their advantages and shortcomings in comparison with other predictive methods, and discuss their conceptual limitations in light of the complexity of biological systems. Furthermore, we provide a survey of published tools for computing PRS and associated resources. The various tools and software packages are categorized based on their technical utility for users or prospective developers. Understanding the array of available tools and their limitations is crucial for accurately assessing and predicting disease risks, facilitating early interventions, and guiding personalized healthcare decisions. Additionally, we also identify potential new avenues for future bioinformatic analyzes and advancements related to PRS.


Asunto(s)
Predisposición Genética a la Enfermedad , Herencia Multifactorial , Programas Informáticos , Humanos , Biología Computacional/métodos , Estudio de Asociación del Genoma Completo/métodos , Factores de Riesgo , Medición de Riesgo/métodos , Puntuación de Riesgo Genético
19.
Brief Bioinform ; 25(3)2024 Mar 27.
Artículo en Inglés | MEDLINE | ID: mdl-38770720

RESUMEN

The normalization of RNA sequencing data is a primary step for downstream analysis. The most popular method used for the normalization is the trimmed mean of M values (TMM) and DESeq. The TMM tries to trim away extreme log fold changes of the data to normalize the raw read counts based on the remaining non-deferentially expressed genes. However, the major problem with the TMM is that the values of trimming factor M are heuristic. This paper tries to estimate the adaptive value of M in TMM based on Jaeckel's Estimator, and each sample acts as a reference to find the scale factor of each sample. The presented approach is validated on SEQC, MAQC2, MAQC3, PICKRELL and two simulated datasets with two-group and three-group conditions by varying the percentage of differential expression and the number of replicates. The performance of the present approach is compared with various state-of-the-art methods, and it is better in terms of area under the receiver operating characteristic curve and differential expression.


Asunto(s)
RNA-Seq , RNA-Seq/métodos , Humanos , Algoritmos , Análisis de Secuencia de ARN/métodos , Biología Computacional/métodos , Perfilación de la Expresión Génica/métodos , Curva ROC , Programas Informáticos
20.
Sensors (Basel) ; 24(9)2024 Apr 24.
Artículo en Inglés | MEDLINE | ID: mdl-38732794

RESUMEN

High-quality eye-tracking data are crucial in behavioral sciences and medicine. Even with a solid understanding of the literature, selecting the most suitable algorithm for a specific research project poses a challenge. Empowering applied researchers to choose the best-fitting detector for their research needs is the primary contribution of this paper. We developed a framework to systematically assess and compare the effectiveness of 13 state-of-the-art algorithms through a unified application interface. Hence, we more than double the number of algorithms that are currently usable within a single software package and allow researchers to identify the best-suited algorithm for a given scientific setup. Our framework validation on retrospective data underscores its suitability for algorithm selection. Through a detailed and reproducible step-by-step workflow, we hope to contribute towards significantly improved data quality in scientific experiments.


Asunto(s)
Algoritmos , Tecnología de Seguimiento Ocular , Humanos , Programas Informáticos , Exactitud de los Datos , Movimientos Oculares/fisiología , Reproducibilidad de los Resultados
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA