Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 31
Filtrar
1.
Syst Biol ; 72(5): 1136-1153, 2023 11 01.
Artigo em Inglês | MEDLINE | ID: mdl-37458991

RESUMO

Divergence time estimation is crucial to provide temporal signals for dating biologically important events from species divergence to viral transmissions in space and time. With the advent of high-throughput sequencing, recent Bayesian phylogenetic studies have analyzed hundreds to thousands of sequences. Such large-scale analyses challenge divergence time reconstruction by requiring inference on highly correlated internal node heights that often become computationally infeasible. To overcome this limitation, we explore a ratio transformation that maps the original $N-1$ internal node heights into a space of one height parameter and $N-2$ ratio parameters. To make the analyses scalable, we develop a collection of linear-time algorithms to compute the gradient and Jacobian-associated terms of the log-likelihood with respect to these ratios. We then apply Hamiltonian Monte Carlo sampling with the ratio transform in a Bayesian framework to learn the divergence times in 4 pathogenic viruses (West Nile virus, rabies virus, Lassa virus, and Ebola virus) and the coralline red algae. Our method both resolves a mixing issue in the West Nile virus example and improves inference efficiency by at least 5-fold for the Lassa and rabies virus examples as well as for the algae example. Our method now also makes it computationally feasible to incorporate mixed-effects molecular clock models for the Ebola virus example, confirms the findings from the original study, and reveals clearer multimodal distributions of the divergence times of some clades of interest.


Assuntos
Algoritmos , Filogenia , Teorema de Bayes , Fatores de Tempo , Método de Monte Carlo
2.
Mol Ecol ; 32(20): 5528-5540, 2023 10.
Artigo em Inglês | MEDLINE | ID: mdl-37706673

RESUMO

Understanding the geographic linkages among populations across the annual cycle is an essential component for understanding the ecology and evolution of migratory species and for facilitating their effective conservation. While genetic markers have been widely applied to describe migratory connections, the rapid development of new sequencing methods, such as low-coverage whole genome sequencing (lcWGS), provides new opportunities for improved estimates of migratory connectivity. Here, we use lcWGS to identify fine-scale population structure in a widespread songbird, the American Redstart (Setophaga ruticilla), and accurately assign individuals to genetically distinct breeding populations. Assignment of individuals from the nonbreeding range reveals population-specific patterns of varying migratory connectivity. By combining migratory connectivity results with demographic analysis of population abundance and trends, we consider full annual cycle conservation strategies for preserving numbers of individuals and genetic diversity. Notably, we highlight the importance of the Northern Temperate-Greater Antilles migratory population as containing the largest proportion of individuals in the species. Finally, we highlight valuable considerations for other population assignment studies aimed at using lcWGS. Our results have broad implications for improving our understanding of the ecology and evolution of migratory species through conservation genomics approaches.


Assuntos
Passeriformes , Aves Canoras , Humanos , Animais , Estados Unidos , Migração Animal , Passeriformes/genética , Aves Canoras/genética , Sequenciamento Completo do Genoma , Região do Caribe
3.
BMC Med Res Methodol ; 23(1): 146, 2023 06 21.
Artigo em Inglês | MEDLINE | ID: mdl-37344771

RESUMO

BACKGROUND: Cochran's Q statistic is routinely used for testing heterogeneity in meta-analysis. Its expected value (under an incorrect null distribution) is part of several popular estimators of the between-study variance, [Formula: see text]. Those applications generally do not account for use of the studies' estimated variances in the inverse-variance weights that define Q (more explicitly, [Formula: see text]). Importantly, those weights make approximating the distribution of [Formula: see text] rather complicated. METHODS: As an alternative, we are investigating a Q statistic, [Formula: see text], whose constant weights use only the studies' arm-level sample sizes. For log-odds-ratio (LOR), log-relative-risk (LRR), and risk difference (RD) as the measures of effect, we study, by simulation, approximations to distributions of [Formula: see text] and [Formula: see text], as the basis for tests of heterogeneity. RESULTS: The results show that: for LOR and LRR, a two-moment gamma approximation to the distribution of [Formula: see text] works well for small sample sizes, and an approximation based on an algorithm of Farebrother is recommended for larger sample sizes. For RD, the Farebrother approximation works very well, even for small sample sizes. For [Formula: see text], the standard chi-square approximation provides levels that are much too low for LOR and LRR and too high for RD. The Kulinskaya et al. (Res Synth Methods 2:254-70, 2011) approximation for RD and the Kulinskaya and Dollinger (BMC Med Res Methodol 15:49, 2015) approximation for LOR work well for [Formula: see text] but have some convergence issues for very small sample sizes combined with small probabilities. CONCLUSIONS: The performance of the standard [Formula: see text] approximation is inadequate for all three binary effect measures. Instead, we recommend a test of heterogeneity based on [Formula: see text] and provide practical guidelines for choosing an appropriate test at the .05 level for all three effect measures.


Assuntos
Algoritmos , Humanos , Simulação por Computador , Probabilidade , Razão de Chances , Tamanho da Amostra
4.
Biometrics ; 78(1): 388-398, 2022 03.
Artigo em Inglês | MEDLINE | ID: mdl-33226116

RESUMO

Inverse probability of treatment weights (IPTWs) are commonly used to control for confounding when estimating causal effects of point exposures from observational data. When planning a study that will be analyzed with IPTWs, determining the required sample size for a given level of statistical power is challenging because of the effect of weighting on the variance of the estimated causal means. This paper considers the utility of the design effect to quantify the effect of weighting on the precision of causal estimates. The design effect is defined as the ratio of the variance of the causal mean estimator divided by the variance of a naïve estimator if, counter to fact, no confounding had been present and weights were not needed. A simple, closed-form approximation of the design effect is derived that is outcome invariant and can be estimated during the study design phase. Once the design effect is approximated for each treatment group, sample size calculations are conducted as for a randomized trial, but with variances inflated by the design effects to account for weighting. Simulations demonstrate the accuracy of the design effect approximation, and practical considerations are discussed.


Assuntos
Modelos Estatísticos , Projetos de Pesquisa , Causalidade , Probabilidade , Tamanho da Amostra
5.
Stat Med ; 40(22): 4794-4808, 2021 09 30.
Artigo em Inglês | MEDLINE | ID: mdl-34126656

RESUMO

As the availability of real-world data sources (eg, EHRs, claims data, registries) and historical data has rapidly surged in recent years, there is an increasing interest and need from investigators and health authorities to leverage all available information to reduce patient burden and accelerate both drug development and regulatory decision making. Bayesian meta-analytic approaches are a popular historical borrowing method that has been developed to leverage such data using robust hierarchical models. The model structure accounts for various degrees of between-trial heterogeneity, resulting in adaptively discounting the external information in the case of data conflict. In this article, we propose to integrate the propensity score method and Bayesian meta-analytic-predictive (MAP) prior to leverage external real-world and historical data. The propensity score methodology is applied to select a subset of patients from external data that are similar to those in the current study with regards to key baseline covariates and to stratify the selected patients together with those in the current study into more homogeneous strata. The MAP prior approach is used to obtain stratum-specific MAP prior and derive the overall propensity score integrated meta-analytic predictive (PS-MAP) prior. Additionally, we allow for tuning the prior effective sample size for the proposed PS-MAP prior, which quantifies the amount of information borrowed from external data. We evaluate the performance of the proposed PS-MAP prior by comparing it to the existing propensity score-integrated power prior approach in a simulation study and illustrate its implementation with an example of a single-arm phase II trial.


Assuntos
Projetos de Pesquisa , Teorema de Bayes , Simulação por Computador , Humanos , Pontuação de Propensão , Tamanho da Amostra
6.
Biometrics ; 76(1): 326-336, 2020 03.
Artigo em Inglês | MEDLINE | ID: mdl-31364156

RESUMO

Bayesian methods allow borrowing of historical information through prior distributions. The concept of prior effective sample size (prior ESS) facilitates quantification and communication of such prior information by equating it to a sample size. Prior information can arise from historical observations; thus, the traditional approach identifies the ESS with such a historical sample size. However, this measure is independent of newly observed data, and thus would not capture an actual "loss of information" induced by the prior in case of prior-data conflict. We build on a recent work to relate prior impact to the number of (virtual) samples from the current data model and introduce the effective current sample size (ECSS) of a prior, tailored to the application in Bayesian clinical trial designs. Special emphasis is put on robust mixture, power, and commensurate priors. We apply the approach to an adaptive design in which the number of recruited patients is adjusted depending on the effective sample size at an interim analysis. We argue that the ECSS is the appropriate measure in this case, as the aim is to save current (as opposed to historical) patients from recruitment. Furthermore, the ECSS can help overcome lack of consensus in the ESS assessment of mixture priors and can, more broadly, provide further insights into the impact of priors. An R package accompanies the paper.


Assuntos
Ensaios Clínicos Adaptados como Assunto/métodos , Ensaios Clínicos Adaptados como Assunto/estatística & dados numéricos , Biometria/métodos , Modelos Estatísticos , Tamanho da Amostra , Teorema de Bayes , Simulação por Computador , Interpretação Estatística de Dados , Humanos
7.
Stata J ; 19(3): 510-522, 2019 Sep.
Artigo em Inglês | MEDLINE | ID: mdl-31814807

RESUMO

In August 2017 the National Center for Health Statistics (NCHS), part of the U.S. Federal Statistical System, published new standards for determining the reliability of proportions estimated using their data. These standards require an individual to take the Korn-Graubard confidence interval (CI), along with CI widths, sample size, and degrees of freedom, to assess reliability of a proportion and determine if it can be presented. The assessment itself involves determining if several conditions are met. This manuscript presents kg_nchs, a postestimation command that is used following svy: proportion. It allows Stata users to (a) calculate the Korn-Graubard CI and associated statistics used in applying the NCHS presentation standards for proportions, and (b) display a series of three dichotomous flags that show if the standards are met. The empirical examples provided show how kg_nchs can be used to easily apply the standards and prevent Stata users from needing to perform manual calculations. While developed for NCHS survey data, this command can also be used with data that stems from any survey with a complex sample design.

8.
Stat Med ; 37(28): 4114-4125, 2018 12 10.
Artigo em Inglês | MEDLINE | ID: mdl-30019428

RESUMO

Network meta-analysis (NMA) has become an increasingly used tool to compare multiple treatments simultaneously by synthesizing direct and indirect evidence in clinical research. However, many existing studies did not properly report the evidence of treatment comparisons and show the comparison structure to audience. In addition, nearly all treatment networks presented only direct evidence, not overall evidence that can reflect the benefit of performing NMAs. This article classifies treatment networks into three types under different assumptions; they include networks with each treatment comparison's edge width proportional to the corresponding number of studies, sample size, and precision. In addition, three new measures (ie, the effective number of studies, the effective sample size, and the effective precision) are proposed to preliminarily quantify overall evidence gained in NMAs. They permit audience to intuitively evaluate the benefit of performing NMAs, compared with pairwise meta-analyses based on only direct evidence. We use four case studies, including one illustrative example, to demonstrate their derivations and interpretations. Treatment networks may look fairly differently when different measures are used to present the evidence. The proposed measures provide clear information about overall evidence of all treatment comparisons, and they also imply the additional number of studies, sample size, and precision obtained from indirect evidence. Some comparisons may benefit little from NMAs. Researchers are encouraged to present overall evidence of all treatment comparisons, so that audience can preliminarily evaluate the quality of NMAs.


Assuntos
Metanálise em Rede , Estatística como Assunto , Interpretação Estatística de Dados , Humanos , Modelos Estatísticos , Tamanho da Amostra , Resultado do Tratamento
9.
Pharm Stat ; 16(4): 232-249, 2017 07.
Artigo em Inglês | MEDLINE | ID: mdl-28448684

RESUMO

Children represent a large underserved population of "therapeutic orphans," as an estimated 80% of children are treated off-label. However, pediatric drug development often faces substantial challenges, including economic, logistical, technical, and ethical barriers, among others. Among many efforts trying to remove these barriers, increased recent attention has been paid to extrapolation; that is, the leveraging of available data from adults or older age groups to draw conclusions for the pediatric population. The Bayesian statistical paradigm is natural in this setting, as it permits the combining (or "borrowing") of information across disparate sources, such as the adult and pediatric data. In this paper, authored by the pediatric subteam of the Drug Information Association Bayesian Scientific Working Group and Adaptive Design Working Group, we develop, illustrate, and provide suggestions on Bayesian statistical methods that could be used to design improved pediatric development programs that use all available information in the most efficient manner. A variety of relevant Bayesian approaches are described, several of which are illustrated through 2 case studies: extrapolating adult efficacy data to expand the labeling for Remicade to include pediatric ulcerative colitis and extrapolating adult exposure-response information for antiepileptic drugs to pediatrics.


Assuntos
Ensaios Clínicos como Assunto , Adulto , Teorema de Bayes , Colite Ulcerativa , Avaliação de Medicamentos , Humanos , Modelos Estatísticos , Projetos de Pesquisa
10.
J Theor Biol ; 407: 371-386, 2016 10 21.
Artigo em Inglês | MEDLINE | ID: mdl-27343033

RESUMO

In this paper I address the question-how large is a phylogenetic sample? I propose a definition of a phylogenetic effective sample size for Brownian motion and Ornstein-Uhlenbeck processes-the regression effective sample size. I discuss how mutual information can be used to define an effective sample size in the non-normal process case and compare these two definitions to an already present concept of effective sample size (the mean effective sample size). Through a simulation study I find that the AICc is robust if one corrects for the number of species or effective number of species. Lastly I discuss how the concept of the phylogenetic effective sample size can be useful for biodiversity quantification, identification of interesting clades and deciding on the importance of phylogenetic correlations.


Assuntos
Filogenia , Tamanho da Amostra , Animais , Biodiversidade , Simulação por Computador , Sequência Conservada , Fenótipo
11.
J Environ Manage ; 173: 121-6, 2016 May 15.
Artigo em Inglês | MEDLINE | ID: mdl-26985730

RESUMO

The effects of climate change and human activities on grassland degradation and soil carbon stocks have become a focus of both research and policy. However, lack of research on appropriate sampling design prevents accurate assessment of soil carbon stocks and stock changes at community and regional scales. Here, we conducted an intensive survey with 1196 sampling sites over an area of 190 km(2) of degraded alpine meadow. Compared to lightly degraded meadow, soil organic carbon (SOC) stocks in moderately, heavily and extremely degraded meadow were reduced by 11.0%, 13.5% and 17.9%, respectively. Our field survey sampling design was overly intensive to estimate SOC status with a tolerable uncertainty of 10%. Power analysis showed that the optimal sampling density to achieve the desired accuracy would be 2, 3, 5 and 7 sites per 10 km(2) for lightly, moderately, heavily and extremely degraded meadows, respectively. If a subsequent paired sampling design with the optimum sample size were performed, assuming stock change rates predicted by experimental and modeling results, we estimate that about 5-10 years would be necessary to detect expected trends in SOC in the top 20 cm soil layer. Our results highlight the utility of conducting preliminary surveys to estimate the appropriate sampling density and avoid wasting resources due to over-sampling, and to estimate the sampling interval required to detect an expected sequestration rate. Future studies will be needed to evaluate spatial and temporal patterns of SOC variability.


Assuntos
Carbono/análise , Monitoramento Ambiental/métodos , Pradaria , Solo/química , Biodegradação Ambiental , Biomassa , Sequestro de Carbono , Plantas , Tamanho da Amostra , Tibet
12.
Pharm Dev Technol ; 21(2): 147-51, 2016 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-25384711

RESUMO

A risk- and science-based approach to control the quality in pharmaceutical manufacturing includes a full understanding of how product attributes and process parameters relate to product performance through a proactive approach in formulation and process development. For dry manufacturing, where moisture content is not directly manipulated within the process, the variability in moisture of the incoming raw materials can impact both the processability and drug product quality attributes. A statistical approach is developed using individual raw material historical lots as a basis for the calculation of tolerance intervals for drug product moisture content so that risks associated with excursions in moisture content can be mitigated. The proposed method is based on a model-independent approach that uses available data to estimate parameters of interest that describe the population of blend moisture content values and which do not require knowledge of the individual blend moisture content values. Another advantage of the proposed tolerance intervals is that, it does not require the use of tabulated values for tolerance factors. This facilitates the implementation on any spreadsheet program like Microsoft Excel. A computational example is used to demonstrate the proposed method.


Assuntos
Composição de Medicamentos/métodos , Preparações Farmacêuticas/química , Água/química , Química Farmacêutica/métodos , Controle de Qualidade , Gestão de Riscos/métodos
13.
Bayesian Anal ; 19(2): 565-593, 2024 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-38665694

RESUMO

Bayesian inference is a popular and widely-used approach to infer phylogenies (evolutionary trees). However, despite decades of widespread application, it remains difficult to judge how well a given Bayesian Markov chain Monte Carlo (MCMC) run explores the space of phylogenetic trees. In this paper, we investigate the Monte Carlo error of phylogenies, focusing on high-dimensional summaries of the posterior distribution, including variability in estimated edge/branch (known in phylogenetics as "split") probabilities and tree probabilities, and variability in the estimated summary tree. Specifically, we ask if there is any measure of effective sample size (ESS) applicable to phylogenetic trees which is capable of capturing the Monte Carlo error of these three summary measures. We find that there are some ESS measures capable of capturing the error inherent in using MCMC samples to approximate the posterior distributions on phylogenies. We term these tree ESS measures, and identify a set of three which are useful in practice for assessing the Monte Carlo error. Lastly, we present visualization tools that can improve comparisons between multiple independent MCMC runs by accounting for the Monte Carlo error present in each chain. Our results indicate that common post-MCMC workflows are insufficient to capture the inherent Monte Carlo error of the tree, and highlight the need for both within-chain mixing and between-chain convergence assessments.

14.
SoftwareX ; 222023 May.
Artigo em Inglês | MEDLINE | ID: mdl-37377886

RESUMO

Bayesian inference has become an attractive choice for scientists seeking to incorporate prior knowledge into their modeling framework. While the R community has been an important contributor in facilitating Bayesian statistical analyses, software to evaluate the impact of prior knowledge to such modeling framework has been lacking. In this article, we present BayesESS, a comprehensive, free, and open source R package for quantifying the impact of parametric priors in Bayesian analysis. We also introduce an accompanying web-based application for estimating and visualizing Bayesian effective sample size for purposes of conducting or planning Bayesian analyses.

15.
Res Synth Methods ; 14(5): 671-688, 2023 Sep.
Artigo em Inglês | MEDLINE | ID: mdl-37381621

RESUMO

For estimation of heterogeneity variance τ 2 in meta-analysis of log-odds-ratio, we derive new mean- and median-unbiased point estimators and new interval estimators based on a generalized Q statistic, Q F , in which the weights depend on only the studies' effective sample sizes. We compare them with familiar estimators based on the inverse-variance-weights version of Q , Q IV . In an extensive simulation, we studied the bias (including median bias) of the point estimators and the coverage (including left and right coverage error) of the confidence intervals. Most estimators add 0.5 to each cell of the 2 × 2 table when one cell contains a zero count; we include a version that always adds 0.5 . The results show that: two of the new point estimators and two of the familiar point estimators are almost unbiased when the total sample size n ≥ 250 and the probability in the Control arm ( p iC ) is 0.1, and when n ≥ 100 and p iC is 0.2 or 0.5; for 0.1 ≤ τ 2 ≤ 1 , all estimators have negative bias for small to medium sample sizes, but for larger sample sizes some of the new median-unbiased estimators are almost median-unbiased; choices of interval estimators depend on values of parameters, but one of the new estimators is reasonable when p iC = 0.1 and another, when p iC = 0.2 or p iC = 0.5 ; and lack of balance between left and right coverage errors for small n and/or p iC implies that the available approximations for the distributions of Q IV and Q F are accurate only for larger sample sizes.


Assuntos
Razão de Chances , Probabilidade , Simulação por Computador , Tamanho da Amostra , Viés
16.
Biol Psychiatry ; 93(1): 29-36, 2023 01 01.
Artigo em Inglês | MEDLINE | ID: mdl-35973856

RESUMO

BACKGROUND: Single nucleotide polymorphism-based heritability is a fundamental quantity in the genetic analysis of complex traits. For case-control phenotypes, for which the continuous distribution of risk in the population is unobserved, observed-scale heritability estimates must be transformed to the more interpretable liability scale. This article describes how the field standard approach incorrectly performs the liability correction in that it does not appropriately account for variation in the proportion of cases across the cohorts comprising the meta-analysis. We propose a simple solution that incorporates cohort-specific ascertainment using the summation of effective sample sizes across cohorts. This solution is applied at the stage of single nucleotide polymorphism-based heritability estimation and does not require generating updated meta-analytic genome-wide association study summary statistics. METHODS: We began by performing a series of simulations to examine the ability of the standard approach and our proposed approach to recapture liability-scale heritability in the population. We went on to examine the differences in estimates obtained from these 2 approaches for real data for 12 major case-control genome-wide association studies of psychiatric and neurologic traits. RESULTS: We found that the field standard approach for performing the liability conversion can downwardly bias estimates by as much as approximately 50% in simulation and approximately 30% in real data. CONCLUSIONS: Prior estimates of liability-scale heritability for genome-wide association study meta-analysis may be drastically underestimated. To this end, we strongly recommend using our proposed approach of using the sum of effective sample sizes across contributing cohorts to obtain unbiased estimates.


Assuntos
Estudo de Associação Genômica Ampla , Herança Multifatorial , Polimorfismo de Nucleotídeo Único/genética , Fenótipo , Estudos de Casos e Controles
17.
SSM Popul Health ; 21: 101317, 2023 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-36589273

RESUMO

Individuals who share similar socio-economic and cultural characteristics also share similar health outcomes. Consequently, they have a propensity to cluster together, which results in positive intra-class correlation coefficients (ICCs) in their socio-demographic and behavioural characteristics. In this study, using data from four rounds of the National Family Health Survey (NFHS), we estimated the ICC for selected socio-demographic and behavioural characteristics in rural and urban areas of six states namely Assam, Gujarat, Kerala, Punjab, Uttar Pradesh, and West Bengal. The socio-demographic and behavioural characteristics included religion & caste of the household head, use of contraception & prevalence of anaemia among currently married women and coverage of full immunization services among children aged 12-23 months. ICC was computed at the level ofPrimary Sampling Units (PSUs), that is, villages in rural areas and census enumeration blocks in urban areas. Our research highlights high clustering in terms of religion and caste within PSUs in India. In NFHS-4, the ICCs for religion ranged from the lowest of 0.19 in rural areas of Kerala to the highest of 0.67 in urban areas of West Bengal. For the caste of the household head, the ICCs ranged from the lowest of 0.12 in the urban areas of Punjab to the highest of 0.46 in the rural areas of Assam. In most of the states selected for the study, the values of ICC were higher for the use of family planning methods than for full immunization. The value of ICC for use of contraception was highest for rural areas of Assam (0.15) followed by rural areas of Gujarat (0.13). A higher value of ICC has considerable implications for determining an effective sample size for large-scale surveys. Our findings agree with the fact that for a given cluster size, the higher the value of ICC, the higher is the loss in precision of the estimate. Knowing and taking into account ICCs can be extremely helpful in determining an effective sample size when designing a large-scale demographic and health survey to arrive at estimates of parameters with the desired precision.

18.
Int J Pharm ; 615: 121462, 2022 Mar 05.
Artigo em Inglês | MEDLINE | ID: mdl-35026317

RESUMO

Near infrared (NIR) spectroscopy has been widely recognized as a powerful PAT tool for monitoring blend uniformity in continuous manufacturing (CM) processes. However, the dynamic nature of the powder stream and the fast rate at which it moves, compared to batch processes, introduces challenges to NIR quantitative methods for monitoring blend uniformity. For instance, defining the effective sample size interrogated by NIR, selecting the best sampling location for blend monitoring, and ensuring NIR model robustness against influential sources of variability are challenges commonly reported for NIR applications in CM. This article reviews the NIR applications for powder blend monitoring in the continuous manufacturing of solid oral dosage forms, with a particular focus on the challenges, opportunities for method optimization and recent advances with respect three main aspects: effective sample size measured by NIR, probe location and method robustness.


Assuntos
Espectroscopia de Luz Próxima ao Infravermelho , Tecnologia Farmacêutica , Composição de Medicamentos , Pós , Comprimidos
19.
Res Synth Methods ; 13(1): 48-67, 2022 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-34427058

RESUMO

To present time-varying evidence, cumulative meta-analysis (CMA) updates results of previous meta-analyses to incorporate new study results. We investigate the properties of CMA, suggest possible improvements and provide the first in-depth simulation study of the use of CMA and CUSUM methods for detection of temporal trends in random-effects meta-analysis. We use the standardized mean difference (SMD) as an effect measure of interest. For CMA, we compare the standard inverse-variance-weighted estimation of the overall effect using REML-based estimation of between-study variance τ 2 with the sample-size-weighted estimation of the effect accompanied by Kulinskaya-Dollinger-Bjørkestøl (Biometrics. 2011; 67:203-212) (KDB) estimation of τ 2 . For all methods, we consider Type 1 error under no shift and power under a shift in the mean in the random-effects model. To ameliorate the lack of power in CMA, we introduce two-stage CMA, in which τ 2 is estimated at Stage 1 (from the first 5-10 studies), and further CMA monitors a target value of effect, keeping the τ 2 value fixed. We recommend this two-stage CMA combined with cumulative testing for positive shift in τ 2 . In practice, use of CMA requires at least 15-20 studies.


Assuntos
Tamanho da Amostra , Simulação por Computador
20.
G3 (Bethesda) ; 11(6)2021 06 17.
Artigo em Inglês | MEDLINE | ID: mdl-33734375

RESUMO

The effective sample size (ESS) is a metric used to summarize in a single term the amount of correlation in a sample. It is of particular interest when predicting the statistical power of genome-wide association studies (GWAS) based on linear mixed models. Here, we introduce an analytical form of the ESS for mixed-model GWAS of quantitative traits and relate it to empirical estimators recently proposed. Using our framework, we derived approximations of the ESS for analyses of related and unrelated samples and for both marginal genetic and gene-environment interaction tests. We conducted simulations to validate our approximations and to provide a quantitative perspective on the statistical power of various scenarios, including power loss due to family relatedness and power gains due to conditioning on the polygenic signal. Our analyses also demonstrate that the power of gene-environment interaction GWAS in related individuals strongly depends on the family structure and exposure distribution. Finally, we performed a series of mixed-model GWAS on data from the UK Biobank and confirmed the simulation results. We notably found that the expected power drop due to family relatedness in the UK Biobank is negligible.


Assuntos
Estudo de Associação Genômica Ampla , Herança Multifatorial , Humanos , Estudo de Associação Genômica Ampla/métodos , Tamanho da Amostra , Herança Multifatorial/genética , Fenótipo , Modelos Lineares , Polimorfismo de Nucleotídeo Único
SELEÇÃO DE REFERÊNCIAS
Detalhe da pesquisa