Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 89.525
Filtrar
Más filtros

Colección CLAP
Intervalo de año de publicación
1.
Cell ; 184(23): 5699-5714.e11, 2021 11 11.
Artículo en Inglés | MEDLINE | ID: mdl-34735795

RESUMEN

Extension of the interval between vaccine doses for the BNT162b2 mRNA vaccine was introduced in the United Kingdom to accelerate population coverage with a single dose. At this time, trial data were lacking, and we addressed this in a study of United Kingdom healthcare workers. The first vaccine dose induced protection from infection from the circulating alpha (B.1.1.7) variant over several weeks. In a substudy of 589 individuals, we show that this single dose induces severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) neutralizing antibody (NAb) responses and a sustained B and T cell response to the spike protein. NAb levels were higher after the extended dosing interval (6-14 weeks) compared with the conventional 3- to 4-week regimen, accompanied by enrichment of CD4+ T cells expressing interleukin-2 (IL-2). Prior SARS-CoV-2 infection amplified and accelerated the response. These data on dynamic cellular and humoral responses indicate that extension of the dosing interval is an effective immunogenic protocol.


Asunto(s)
Vacunas contra la COVID-19/inmunología , Vacunas Sintéticas/inmunología , Adulto , Anciano , Anticuerpos Neutralizantes/inmunología , Anticuerpos Antivirales/inmunología , Vacuna BNT162 , COVID-19/sangre , COVID-19/inmunología , COVID-19/virología , Reactividad Cruzada/inmunología , Relación Dosis-Respuesta Inmunológica , Etnicidad , Femenino , Humanos , Inmunidad , Inmunoglobulina G/inmunología , Modelos Lineales , Masculino , Persona de Mediana Edad , Estándares de Referencia , SARS-CoV-2/inmunología , Linfocitos T/inmunología , Resultado del Tratamiento , Adulto Joven , Vacunas de ARNm
2.
Cell ; 183(7): 1986-2002.e26, 2020 12 23.
Artículo en Inglés | MEDLINE | ID: mdl-33333022

RESUMEN

Serotonin plays a central role in cognition and is the target of most pharmaceuticals for psychiatric disorders. Existing drugs have limited efficacy; creation of improved versions will require better understanding of serotonergic circuitry, which has been hampered by our inability to monitor serotonin release and transport with high spatial and temporal resolution. We developed and applied a binding-pocket redesign strategy, guided by machine learning, to create a high-performance, soluble, fluorescent serotonin sensor (iSeroSnFR), enabling optical detection of millisecond-scale serotonin transients. We demonstrate that iSeroSnFR can be used to detect serotonin release in freely behaving mice during fear conditioning, social interaction, and sleep/wake transitions. We also developed a robust assay of serotonin transporter function and modulation by drugs. We expect that both machine-learning-guided binding-pocket redesign and iSeroSnFR will have broad utility for the development of other sensors and in vitro and in vivo serotonin detection, respectively.


Asunto(s)
Evolución Molecular Dirigida , Aprendizaje Automático , Serotonina/metabolismo , Algoritmos , Secuencia de Aminoácidos , Amígdala del Cerebelo/fisiología , Animales , Conducta Animal , Sitios de Unión , Encéfalo/metabolismo , Células HEK293 , Humanos , Cinética , Modelos Lineales , Ratones , Ratones Endogámicos C57BL , Fotones , Unión Proteica , Proteínas de Transporte de Serotonina en la Membrana Plasmática/metabolismo , Sueño/fisiología , Vigilia/fisiología
3.
Cell ; 164(1-2): 293-309, 2016 Jan 14.
Artículo en Inglés | MEDLINE | ID: mdl-26771497

RESUMEN

Large-scale genomic studies have identified multiple somatic aberrations in breast cancer, including copy number alterations and point mutations. Still, identifying causal variants and emergent vulnerabilities that arise as a consequence of genetic alterations remain major challenges. We performed whole-genome small hairpin RNA (shRNA) "dropout screens" on 77 breast cancer cell lines. Using a hierarchical linear regression algorithm to score our screen results and integrate them with accompanying detailed genetic and proteomic information, we identify vulnerabilities in breast cancer, including candidate "drivers," and reveal general functional genomic properties of cancer cells. Comparisons of gene essentiality with drug sensitivity data suggest potential resistance mechanisms, effects of existing anti-cancer drugs, and opportunities for combination therapy. Finally, we demonstrate the utility of this large dataset by identifying BRD4 as a potential target in luminal breast cancer and PIK3CA mutations as a resistance determinant for BET-inhibitors.


Asunto(s)
Algoritmos , Neoplasias de la Mama/genética , Neoplasias de la Mama/tratamiento farmacológico , Neoplasias de la Mama/patología , Proteínas de Ciclo Celular , Línea Celular Tumoral , Fosfatidilinositol 3-Quinasa Clase I , Análisis por Conglomerados , Resistencia a Antineoplásicos , Dosificación de Gen , Perfilación de la Expresión Génica , Estudio de Asociación del Genoma Completo , Humanos , Modelos Lineales , Proteínas Nucleares/genética , Fosfatidilinositol 3-Quinasas , Factores de Transcripción/genética
4.
Nature ; 634(8032): 201-209, 2024 Oct.
Artículo en Inglés | MEDLINE | ID: mdl-39358526

RESUMEN

A goal of neuroscience is to obtain a causal model of the nervous system. The recently reported whole-brain fly connectome1-3 specifies the synaptic paths by which neurons can affect each other, but not how strongly they do affect each other in vivo. To overcome this limitation, we introduce a combined experimental and statistical strategy for efficiently learning a causal model of the fly brain, which we refer to as the 'effectome'. Specifically, we propose an estimator for a linear dynamical model of the fly brain that uses stochastic optogenetic perturbation data to estimate causal effects and the connectome as a prior to greatly improve estimation efficiency. We validate our estimator in connectome-based linear simulations and show that it recovers a linear approximation to the nonlinear dynamics of more biophysically realistic simulations. We then analyse the connectome to propose circuits that dominate the dynamics of the fly nervous system. We discover that the dominant circuits involve only relatively small populations of neurons-thus, neuron-level imaging, stimulation and identification are feasible. This approach also re-discovers known circuits and generates testable hypotheses about their dynamics. Overall, we provide evidence that fly whole-brain dynamics are generated by a large collection of small circuits that operate largely independently of each other. This implies that a causal model of a brain can be feasibly obtained in the fly.


Asunto(s)
Encéfalo , Conectoma , Drosophila melanogaster , Vías Nerviosas , Neuronas , Animales , Femenino , Encéfalo/anatomía & histología , Encéfalo/citología , Encéfalo/fisiología , Drosophila melanogaster/anatomía & histología , Drosophila melanogaster/citología , Drosophila melanogaster/fisiología , Modelos Lineales , Modelos Neurológicos , Neuronas/citología , Neuronas/fisiología , Optogenética , Reproducibilidad de los Resultados , Procesos Estocásticos , Vías Nerviosas/anatomía & histología , Vías Nerviosas/citología , Vías Nerviosas/fisiología
5.
Mol Cell ; 80(2): 359-373.e8, 2020 10 15.
Artículo en Inglés | MEDLINE | ID: mdl-32991830

RESUMEN

Eukaryotic gene expression regulation involves thousands of distal regulatory elements. Understanding the quantitative contribution of individual enhancers to gene expression is critical for assessing the role of disease-associated genetic risk variants. Yet, we lack the ability to accurately link genes with their distal regulatory elements. To address this, we used 3D enhancer-promoter (E-P) associations identified using split-pool recognition of interactions by tag extension (SPRITE) to build a predictive model of gene expression. Our model dramatically outperforms models using genomic proximity and can be used to determine the quantitative impact of enhancer loss on gene expression in different genetic backgrounds. We show that genes that form stable E-P hubs have less cell-to-cell variability in gene expression. Finally, we identified transcription factors that regulate stimulation-dependent E-P interactions. Together, our results provide a framework for understanding quantitative contributions of E-P interactions and associated genetic variants to gene expression.


Asunto(s)
Bacterias/aislamiento & purificación , Elementos de Facilitación Genéticos , Regiones Promotoras Genéticas , Animales , Células Dendríticas/metabolismo , Femenino , Regulación de la Expresión Génica , Modelos Lineales , Ratones Endogámicos C57BL , Modelos Biológicos , Procesos Estocásticos , Factores de Transcripción/metabolismo
6.
Genome Res ; 34(9): 1304-1311, 2024 Oct 11.
Artículo en Inglés | MEDLINE | ID: mdl-39231610

RESUMEN

Linear mixed models (LMMs) have been widely used in genome-wide association studies to control for population stratification and cryptic relatedness. However, estimating LMM parameters is computationally expensive, necessitating large-scale matrix operations to build the genetic relationship matrix (GRM). Over the past 25 years, Randomized Linear Algebra has provided alternative approaches to such matrix operations by leveraging matrix sketching, which often results in provably accurate fast and efficient approximations. We leverage matrix sketching to develop a fast and efficient LMM method called Matrix-Sketching LMM (MaSk-LMM) by sketching the genotype matrix to reduce its dimensions and speed up computations. Our framework comes with both theoretical guarantees and a strong empirical performance compared to the current state-of-the-art for simulated traits and complex diseases.


Asunto(s)
Estudio de Asociación del Genoma Completo , Humanos , Estudio de Asociación del Genoma Completo/métodos , Modelos Lineales , Modelos Genéticos , Genotipo , Algoritmos , Simulación por Computador
7.
Nature ; 592(7855): 571-576, 2021 04.
Artículo en Inglés | MEDLINE | ID: mdl-33790468

RESUMEN

Biological invasions are responsible for substantial biodiversity declines as well as high economic losses to society and monetary expenditures associated with the management of these invasions1,2. The InvaCost database has enabled the generation of a reliable, comprehensive, standardized and easily updatable synthesis of the monetary costs of biological invasions worldwide3. Here we found that the total reported costs of invasions reached a minimum of US$1.288 trillion (2017 US dollars) over the past few decades (1970-2017), with an annual mean cost of US$26.8 billion. Moreover, we estimate that the annual mean cost could reach US$162.7 billion in 2017. These costs remain strongly underestimated and do not show any sign of slowing down, exhibiting a consistent threefold increase per decade. We show that the documented costs are widely distributed and have strong gaps at regional and taxonomic scales, with damage costs being an order of magnitude higher than management expenditures. Research approaches that document the costs of biological invasions need to be further improved. Nonetheless, our findings call for the implementation of consistent management actions and international policy agreements that aim to reduce the burden of invasive alien species.


Asunto(s)
Biodiversidad , Ecología/economía , Ciencia Ambiental/economía , Internacionalidad , Especies Introducidas/economía , Especies Introducidas/tendencias , Animales , Mapeo Geográfico , Invertebrados , Modelos Lineales , Plantas , Vertebrados
8.
Brief Bioinform ; 25(2)2024 Jan 22.
Artículo en Inglés | MEDLINE | ID: mdl-38436558

RESUMEN

Recently, there has been a growing interest in variable selection for causal inference within the context of high-dimensional data. However, when the outcome exhibits a skewed distribution, ensuring the accuracy of variable selection and causal effect estimation might be challenging. Here, we introduce the generalized median adaptive lasso (GMAL) for covariate selection to achieve an accurate estimation of causal effect even when the outcome follows skewed distributions. A distinctive feature of our proposed method is that we utilize a linear median regression model for constructing penalty weights, thereby maintaining the accuracy of variable selection and causal effect estimation even when the outcome presents extremely skewed distributions. Simulation results showed that our proposed method performs comparably to existing methods in variable selection when the outcome follows a symmetric distribution. Besides, the proposed method exhibited obvious superiority over the existing methods when the outcome follows a skewed distribution. Meanwhile, our proposed method consistently outperformed the existing methods in causal estimation, as indicated by smaller root-mean-square error. We also utilized the GMAL method on a deoxyribonucleic acid methylation dataset from the Alzheimer's disease (AD) neuroimaging initiative database to investigate the association between cerebrospinal fluid tau protein levels and the severity of AD.


Asunto(s)
Enfermedad de Alzheimer , Humanos , Enfermedad de Alzheimer/genética , Simulación por Computador , Bases de Datos Factuales , Modelos Lineales , Procesamiento Proteico-Postraduccional
9.
PLoS Genet ; 19(11): e1011022, 2023 Nov.
Artículo en Inglés | MEDLINE | ID: mdl-37934796

RESUMEN

Epigenetic researchers often evaluate DNA methylation as a potential mediator of the effect of social/environmental exposures on a health outcome. Modern statistical methods for jointly evaluating many mediators have not been widely adopted. We compare seven methods for high-dimensional mediation analysis with continuous outcomes through both diverse simulations and analysis of DNAm data from a large multi-ethnic cohort in the United States, while providing an R package for their seamless implementation and adoption. Among the considered choices, the best-performing methods for detecting active mediators in simulations are the Bayesian sparse linear mixed model (BSLMM) and high-dimensional mediation analysis (HDMA); while the preferred methods for estimating the global mediation effect are high-dimensional linear mediation analysis (HILMA) and principal component mediation analysis (PCMA). We provide guidelines for epigenetic researchers on choosing the best method in practice and offer suggestions for future methodological development.


Asunto(s)
Metilación de ADN , Análisis de Mediación , Humanos , Metilación de ADN/genética , Teorema de Bayes , Modelos Lineales , Exposición a Riesgos Ambientales
10.
J Neurosci ; 44(14)2024 Apr 03.
Artículo en Inglés | MEDLINE | ID: mdl-38316565

RESUMEN

Although we must prioritize the processing of task-relevant information to navigate life, our ability to do so fluctuates across time. Previous work has identified fMRI functional connectivity (FC) networks that predict an individual's ability to sustain attention and vary with attentional state from 1 min to the next. However, traditional dynamic FC approaches typically lack the temporal precision to capture moment-to-moment network fluctuations. Recently, researchers have "unfurled" traditional FC matrices in "edge cofluctuation time series" which measure timepoint-by-timepoint cofluctuations between regions. Here we apply event-based and parametric fMRI analyses to edge time series to capture moment-to-moment fluctuations in networks related to attention. In two independent fMRI datasets examining young adults of both sexes in which participants performed a sustained attention task, we identified a reliable set of edges that rapidly deflects in response to rare task events. Another set of edges varies with continuous fluctuations in attention and overlaps with a previously defined set of edges associated with individual differences in sustained attention. Demonstrating that edge-based analyses are not simply redundant with traditional regions-of-interest-based approaches, up to one-third of reliably deflected edges were not predicted from univariate activity patterns alone. These results reveal the large potential in combining traditional fMRI analyses with edge time series to identify rapid reconfigurations in networks across the brain.


Asunto(s)
Atención , Encéfalo , Masculino , Femenino , Adulto Joven , Humanos , Modelos Lineales , Encéfalo/diagnóstico por imagen , Encéfalo/fisiología , Atención/fisiología , Mapeo Encefálico/métodos , Imagen por Resonancia Magnética/métodos
11.
Genet Epidemiol ; 48(4): 164-189, 2024 06.
Artículo en Inglés | MEDLINE | ID: mdl-38420714

RESUMEN

Gene-environment (GxE) interactions play a crucial role in understanding the complex etiology of various traits, but assessing them using observational data can be challenging due to unmeasured confounders for lifestyle and environmental risk factors. Mendelian randomization (MR) has emerged as a valuable method for assessing causal relationships based on observational data. This approach utilizes genetic variants as instrumental variables (IVs) with the aim of providing a valid statistical test and estimation of causal effects in the presence of unmeasured confounders. MR has gained substantial popularity in recent years largely due to the success of genome-wide association studies. Many methods have been developed for MR; however, limited work has been done on evaluating GxE interaction. In this paper, we focus on two primary IV approaches: the two-stage predictor substitution and the two-stage residual inclusion, and extend them to accommodate GxE interaction under both the linear and logistic regression models for continuous and binary outcomes, respectively. Comprehensive simulation study and analytical derivations reveal that resolving the linear regression model is relatively straightforward. In contrast, the logistic regression model presents a considerably more intricate challenge, which demands additional effort.


Asunto(s)
Interacción Gen-Ambiente , Estudio de Asociación del Genoma Completo , Análisis de la Aleatorización Mendeliana , Humanos , Modelos Logísticos , Modelos Lineales , Polimorfismo de Nucleótido Simple , Modelos Genéticos , Variación Genética , Simulación por Computador
12.
Brief Bioinform ; 24(1)2023 01 19.
Artículo en Inglés | MEDLINE | ID: mdl-36545787

RESUMEN

Genotype-by-environment interaction (GEI or GxE) plays an important role in understanding complex human traits. However, it is usually challenging to detect GEI signals efficiently and accurately while adjusting for population stratification and sample relatedness in large-scale genome-wide association studies (GWAS). Here we propose a fast and powerful linear mixed model-based approach, fastGWA-GE, to test for GEI effect and G + GxE joint effect. Our extensive simulations show that fastGWA-GE outperforms other existing GEI test methods by controlling genomic inflation better, providing larger power and running hundreds to thousands of times faster. We performed a fastGWA-GE analysis of ~7.27 million variants on 452 249 individuals of European ancestry for 13 quantitative traits and five environment variables in the UK Biobank GWAS data and identified 96 significant signals (72 variants across 57 loci) with GEI test P-values < 1 × 10-9, including 27 novel GEI associations, which highlights the effectiveness of fastGWA-GE in GEI signal discovery in large-scale GWAS.


Asunto(s)
Interacción Gen-Ambiente , Estudio de Asociación del Genoma Completo , Humanos , Fenotipo , Genotipo , Modelos Lineales , Polimorfismo de Nucleótido Simple
13.
Brief Bioinform ; 24(1)2023 01 19.
Artículo en Inglés | MEDLINE | ID: mdl-36617187

RESUMEN

Differential abundance analysis (DAA) is one central statistical task in microbiome data analysis. A robust and powerful DAA tool can help identify highly confident microbial candidates for further biological validation. Current microbiome studies frequently generate correlated samples from different microbiome sampling schemes such as spatial and temporal sampling. In the past decade, a number of DAA tools for correlated microbiome data (DAA-c) have been proposed. Disturbingly, different DAA-c tools could sometimes produce quite discordant results. To recommend the best practice to the field, we performed the first comprehensive evaluation of existing DAA-c tools using real data-based simulations. Overall, the linear model-based methods LinDA, MaAsLin2 and LDM are more robust than methods based on generalized linear models. The LinDA method is the only method that maintains reasonable performance in the presence of strong compositional effects.


Asunto(s)
Benchmarking , Microbiota , Microbiota/genética , Modelos Lineales , Bases de Datos Factuales , Metagenómica/métodos
14.
Brief Bioinform ; 24(3)2023 05 19.
Artículo en Inglés | MEDLINE | ID: mdl-37141142

RESUMEN

In genome assembly, scaffolding can obtain more complete and continuous scaffolds. Current scaffolding methods usually adopt one type of read to construct a scaffold graph and then orient and order contigs. However, scaffolding with the strengths of two or more types of reads seems to be a better solution to some tricky problems. Combining the advantages of different types of data is significant for scaffolding. Here, a hybrid scaffolding method (SLHSD) is present that simultaneously leverages the precision of short reads and the length advantage of long reads. Building an optimal scaffold graph is an important foundation for getting scaffolds. SLHSD uses a new algorithm that combines long and short read alignment information to determine whether to add an edge and how to calculate the edge weight in a scaffold graph. In addition, SLHSD develops a strategy to ensure that edges with high confidence can be added to the graph with priority. Then, a linear programming model is used to detect and remove remaining false edges in the graph. We compared SLHSD with other scaffolding methods on five datasets. Experimental results show that SLHSD outperforms other methods. The open-source code of SLHSD is available at https://github.com/luojunwei/SLHSD.


Asunto(s)
Algoritmos , Secuenciación de Nucleótidos de Alto Rendimiento , Análisis de Secuencia de ADN/métodos , Secuenciación de Nucleótidos de Alto Rendimiento/métodos , Programas Informáticos , Modelos Lineales
15.
Bioinformatics ; 40(9)2024 Sep 02.
Artículo en Inglés | MEDLINE | ID: mdl-39254590

RESUMEN

MOTIVATION: The functions of genes in networks are typically correlated due to their functional connectivity. Variable selection methods have been developed to select important genes associated with a trait while incorporating network graphical information. However, no method has been proposed to quantify the uncertainty of individual genes under such settings. RESULTS: In this paper, we construct confidence intervals (CIs) and provide P-values for parameters of a high-dimensional linear model incorporating graphical structures where the number of variables p diverges with the number of observations. For combining the graphical information, we propose a graph-constrained desparsified LASSO (least absolute shrinkage and selection operator) (GCDL) estimator, which reduces dramatically the influence of high correlation of predictors and enjoys the advantage of faster computation and higher accuracy compared with the desparsified LASSO. Theoretical results show that the GCDL estimator achieves asymptotic normality. The asymptotic property of the uniform convergence is established, with which an explicit expression of the uniform CI can be derived. Extensive numerical results indicate that the GCDL estimator and its (uniform) CI perform well even when predictors are highly correlated. AVAILABILITY AND IMPLEMENTATION: An R package implementing the proposed method is available at https://github.com/XiaoZhangryy/gcdl.


Asunto(s)
Algoritmos , Modelos Lineales , Incertidumbre , Redes Reguladoras de Genes , Biología Computacional/métodos , Humanos , Perfilación de la Expresión Génica/métodos
16.
Bioinformatics ; 40(8)2024 08 02.
Artículo en Inglés | MEDLINE | ID: mdl-39115884

RESUMEN

MOTIVATION: Generalized linear mixed models (GLMMs), such as the negative-binomial or Poisson linear mixed model, are widely applied to single-cell RNA sequencing data to compare transcript expression between different conditions determined at the subject level. However, the model is computationally intensive, and its relative statistical performance to pseudobulk approaches is poorly understood. RESULTS: We propose offset-pseudobulk as a lightweight alternative to GLMMs. We prove that a count-based pseudobulk equipped with a proper offset variable has the same statistical properties as GLMMs in terms of both point estimates and standard errors. We confirm our findings using simulations based on real data. Offset-pseudobulk is substantially faster (>×10) and numerically more stable than GLMMs. AVAILABILITY AND IMPLEMENTATION: Offset pseudobulk can be easily implemented in any generalized linear model software by tweaking a few options. The codes can be found at https://github.com/hanbin973/pseudobulk_is_mm.


Asunto(s)
Análisis de la Célula Individual , Análisis de la Célula Individual/métodos , Modelos Lineales , Programas Informáticos , Estudios de Casos y Controles , Análisis de Secuencia de ARN/métodos , Humanos , Algoritmos
17.
PLoS Biol ; 20(2): e3001562, 2022 02.
Artículo en Inglés | MEDLINE | ID: mdl-35180228

RESUMEN

The power of language to modify the reader's perception of interpreting biomedical results cannot be underestimated. Misreporting and misinterpretation are pressing problems in randomized controlled trials (RCT) output. This may be partially related to the statistical significance paradigm used in clinical trials centered around a P value below 0.05 cutoff. Strict use of this P value may lead to strategies of clinical researchers to describe their clinical results with P values approaching but not reaching the threshold to be "almost significant." The question is how phrases expressing nonsignificant results have been reported in RCTs over the past 30 years. To this end, we conducted a quantitative analysis of English full texts containing 567,758 RCTs recorded in PubMed between 1990 and 2020 (81.5% of all published RCTs in PubMed). We determined the exact presence of 505 predefined phrases denoting results that approach but do not cross the line of formal statistical significance (P < 0.05). We modeled temporal trends in phrase data with Bayesian linear regression. Evidence for temporal change was obtained through Bayes factor (BF) analysis. In a randomly sampled subset, the associated P values were manually extracted. We identified 61,741 phrases in 49,134 RCTs indicating almost significant results (8.65%; 95% confidence interval (CI): 8.58% to 8.73%). The overall prevalence of these phrases remained stable over time, with the most prevalent phrases being "marginally significant" (in 7,735 RCTs), "all but significant" (7,015), "a nonsignificant trend" (3,442), "failed to reach statistical significance" (2,578), and "a strong trend" (1,700). The strongest evidence for an increased temporal prevalence was found for "a numerical trend," "a positive trend," "an increasing trend," and "nominally significant." In contrast, the phrases "all but significant," "approaches statistical significance," "did not quite reach statistical significance," "difference was apparent," "failed to reach statistical significance," and "not quite significant" decreased over time. In a random sampled subset of 29,000 phrases, the manually identified and corresponding 11,926 P values, 68,1% ranged between 0.05 and 0.15 (CI: 67. to 69.0; median 0.06). Our results show that RCT reports regularly contain specific phrases describing marginally nonsignificant results to report P values close to but above the dominant 0.05 cutoff. The fact that the prevalence of the phrases remained stable over time indicates that this practice of broadly interpreting P values close to a predefined threshold remains prevalent. To enhance responsible and transparent interpretation of RCT results, researchers, clinicians, reviewers, and editors may reduce the focus on formal statistical significance thresholds and stimulate reporting of P values with corresponding effect sizes and CIs and focus on the clinical relevance of the statistical difference found in RCTs.


Asunto(s)
PubMed/normas , Publicaciones/normas , Ensayos Clínicos Controlados Aleatorios como Asunto/normas , Proyectos de Investigación/normas , Informe de Investigación/normas , Teorema de Bayes , Sesgo , Humanos , Modelos Lineales , Evaluación de Resultado en la Atención de Salud/métodos , Evaluación de Resultado en la Atención de Salud/normas , Evaluación de Resultado en la Atención de Salud/estadística & datos numéricos , PubMed/estadística & datos numéricos , Publicaciones/estadística & datos numéricos , Ensayos Clínicos Controlados Aleatorios como Asunto/estadística & datos numéricos , Reproducibilidad de los Resultados
18.
PLoS Comput Biol ; 20(7): e1012142, 2024 Jul.
Artículo en Inglés | MEDLINE | ID: mdl-39047024

RESUMEN

Increasing genetic and phenotypic data size is critical for understanding the genetic determinants of diseases. Evidently, establishing practical means for collaboration and data sharing among institutions is a fundamental methodological barrier for performing high-powered studies. As the sample sizes become more heterogeneous, complex statistical approaches, such as generalized linear mixed effects models, must be used to correct for the confounders that may bias results. On another front, due to the privacy concerns around Protected Health Information (PHI), genetic information is restrictively protected by sharing according to regulations such as Health Insurance Portability and Accountability Act (HIPAA). This limits data sharing among institutions and hampers efforts around executing high-powered collaborative studies. Federated approaches are promising to alleviate the issues around privacy and performance, since sensitive data never leaves the local sites. Motivated by these, we developed FedGMMAT, a federated genetic association testing tool that utilizes a federated statistical testing approach for efficient association tests that can correct for confounding fixed and additive polygenic random effects among different collaborating sites. Genetic data is never shared among collaborating sites, and the intermediate statistics are protected by encryption. Using simulated and real datasets, we demonstrate FedGMMAT can achieve the virtually same results as pooled analysis under a privacy-preserving framework with practical resource requirements.


Asunto(s)
Difusión de la Información , Humanos , Modelos Lineales , Difusión de la Información/métodos , Biología Computacional/métodos , Programas Informáticos , Estudio de Asociación del Genoma Completo/métodos , Estudios de Asociación Genética
19.
PLoS Comput Biol ; 20(4): e1011975, 2024 Apr.
Artículo en Inglés | MEDLINE | ID: mdl-38669271

RESUMEN

The brain produces diverse functions, from perceiving sounds to producing arm reaches, through the collective activity of populations of many neurons. Determining if and how the features of these exogenous variables (e.g., sound frequency, reach angle) are reflected in population neural activity is important for understanding how the brain operates. Often, high-dimensional neural population activity is confined to low-dimensional latent spaces. However, many current methods fail to extract latent spaces that are clearly structured by exogenous variables. This has contributed to a debate about whether or not brains should be thought of as dynamical systems or representational systems. Here, we developed a new latent process Bayesian regression framework, the orthogonal stochastic linear mixing model (OSLMM) which introduces an orthogonality constraint amongst time-varying mixture coefficients, and provide Markov chain Monte Carlo inference procedures. We demonstrate superior performance of OSLMM on latent trajectory recovery in synthetic experiments and show superior computational efficiency and prediction performance on several real-world benchmark data sets. We primarily focus on demonstrating the utility of OSLMM in two neural data sets: µECoG recordings from rat auditory cortex during presentation of pure tones and multi-single unit recordings form monkey motor cortex during complex arm reaching. We show that OSLMM achieves superior or comparable predictive accuracy of neural data and decoding of external variables (e.g., reach velocity). Most importantly, in both experimental contexts, we demonstrate that OSLMM latent trajectories directly reflect features of the sounds and reaches, demonstrating that neural dynamics are structured by neural representations. Together, these results demonstrate that OSLMM will be useful for the analysis of diverse, large-scale biological time-series datasets.


Asunto(s)
Corteza Auditiva , Teorema de Bayes , Cadenas de Markov , Modelos Neurológicos , Neuronas , Procesos Estocásticos , Animales , Ratas , Corteza Auditiva/fisiología , Neuronas/fisiología , Biología Computacional , Modelos Lineales , Método de Montecarlo , Simulación por Computador
20.
Nature ; 568(7751): 221-225, 2019 04.
Artículo en Inglés | MEDLINE | ID: mdl-30944480

RESUMEN

The global land and ocean carbon sinks have increased proportionally with increasing carbon dioxide emissions during the past decades1. It is thought that Northern Hemisphere lands make a dominant contribution to the global land carbon sink2-7; however, the long-term trend of the northern land sink remains uncertain. Here, using measurements of the interhemispheric gradient of atmospheric carbon dioxide from 1958 to 2016, we show that the northern land sink remained stable between the 1960s and the late 1980s, then increased by 0.5 ± 0.4 petagrams of carbon per year during the 1990s and by 0.6 ± 0.5 petagrams of carbon per year during the 2000s. The increase of the northern land sink in the 1990s accounts for 65% of the increase in the global land carbon flux during that period. The subsequent increase in the 2000s is larger than the increase in the global land carbon flux, suggesting a coincident decrease of carbon uptake in the Southern Hemisphere. Comparison of our findings with the simulations of an ensemble of terrestrial carbon models5,8 over the same period suggests that the decadal change in the northern land sink between the 1960s and the 1990s can be explained by a combination of increasing concentrations of atmospheric carbon dioxide, climate variability and changes in land cover. However, the increase during the 2000s is underestimated by all models, which suggests the need for improved consideration of changes in drivers such as nitrogen deposition, diffuse light and land-use change. Overall, our findings underscore the importance of Northern Hemispheric land as a carbon sink.


Asunto(s)
Dióxido de Carbono/análisis , Dióxido de Carbono/historia , Secuestro de Carbono , Mapeo Geográfico , Sedimentos Geológicos/química , Atmósfera/química , Carbono/química , Dióxido de Carbono/química , China , Materiales de Construcción/análisis , Bosques , Combustibles Fósiles/análisis , Historia del Siglo XX , Historia del Siglo XXI , Modelos Lineales , Modelos Teóricos , Nitrógeno/química , Siberia , Incertidumbre
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA