RESUMO
Considerable uncertainty surrounds the timeline of introductions and onsets of local transmission of severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) globally1-7. Although a limited number of SARS-CoV-2 introductions were reported in January and February 2020 (refs.8,9), the narrowness of the initial testing criteria, combined with a slow growth in testing capacity and porous travel screening10, left many countries vulnerable to unmitigated, cryptic transmission. Here we use a global metapopulation epidemic model to provide a mechanistic understanding of the early dispersal of infections and the temporal windows of the introduction of SARS-CoV-2 and onset of local transmission in Europe and the USA. We find that community transmission of SARS-CoV-2 was likely to have been present in several areas of Europe and the USA by January 2020, and estimate that by early March, only 1 to 4 in 100 SARS-CoV-2 infections were detected by surveillance systems. The modelling results highlight international travel as the key driver of the introduction of SARS-CoV-2, with possible introductions and transmission events as early as December 2019 to January 2020. We find a heterogeneous geographic distribution of cumulative infection attack rates by 4 July 2020, ranging from 0.78% to 15.2% across US states and 0.19% to 13.2% in European countries. Our approach complements phylogenetic analyses and other surveillance approaches and provides insights that can be used to design innovative, model-driven surveillance systems that guide enhanced testing and response strategies.
Assuntos
COVID-19/epidemiologia , COVID-19/transmissão , Modelos Epidemiológicos , SARS-CoV-2/isolamento & purificação , Viagem Aérea/estatística & dados numéricos , COVID-19/mortalidade , COVID-19/virologia , China/epidemiologia , Surtos de Doenças/estatística & dados numéricos , Europa (Continente)/epidemiologia , Humanos , Densidade Demográfica , Fatores de Tempo , Estados Unidos/epidemiologiaRESUMO
Detailed characterization of severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) transmission across different settings can help design less disruptive interventions. We used real-time, privacy-enhanced mobility data in the New York City, NY and Seattle, WA metropolitan areas to build a detailed agent-based model of SARS-CoV-2 infection to estimate the where, when, and magnitude of transmission events during the pandemic's first wave. We estimate that only 18% of individuals produce most infections (80%), with about 10% of events that can be considered superspreading events (SSEs). Although mass gatherings present an important risk for SSEs, we estimate that the bulk of transmission occurred in smaller events in settings like workplaces, grocery stores, or food venues. The places most important for transmission change during the pandemic and are different across cities, signaling the large underlying behavioral component underneath them. Our modeling complements case studies and epidemiological data and indicates that real-time tracking of transmission events could help evaluate and define targeted mitigation policies.
Assuntos
COVID-19 , Busca de Comunicante , SARS-CoV-2 , COVID-19/transmissão , Humanos , Cidade de Nova Iorque/epidemiologia , Pandemias , Dinâmica Populacional , Fatores de Tempo , Washington/epidemiologiaRESUMO
Test-negative designs are increasingly used to evaluate vaccine effectiveness because of desirable properties like reduced confounding due to healthcare-seeking behaviors and lower cost compared to other study designs. An individual's decision to seek care often depends on their disease severity, with severe disease more likely to be captured than mild disease. As many vaccines likely attenuate disease severity, this phenomenon generally results in an upward-biased estimate of vaccine effectiveness against symptomatic disease. To address the resulting bias, analytic solutions like adjusting for or matching on severity have been suggested. In this paper, we examine the performance of the test-negative design under different vaccine effects on disease severity and the utility of adjusting or matching on severity. We further consider the implications of studies that focus only on milder disease by restricting recruitment to outpatient settings. Through an analytic framework and simulations accompanied by a real-world example, we demonstrate that, when vaccination attenuates disease severity, the magnitude of bias is influenced by the degree of under-ascertainment of mild disease relative to severe disease. When vaccination does not attenuate disease severity, bias is not present. We further show that analytic fixes negligibly impact bias and that outpatient-only studies frequently produce downward-biased estimates.
RESUMO
Both individually and cluster randomized study designs have been used for vaccine trials to assess the effects of vaccine on reducing the risk of disease or infection. The choice between individually and cluster randomized designs is often driven by the target estimand of interest (eg, direct versus total), statistical power, and, importantly, logistic feasibility. To combat emerging infectious disease threats, especially when the number of events from one single trial may not be adequate to obtain vaccine effect estimates with a desired level of precision, it may be necessary to combine information across multiple trials. In this article, we propose a model formulation to estimate the direct, indirect, total, and overall vaccine effects combining data from trials with two types of study designs: individual-randomization and cluster-randomization, based on a Cox proportional hazards model, where the hazard of infection depends on both vaccine status of the individual as well as the vaccine status of the other individuals in the same cluster. We illustrate the use of the proposed model and assess the potential efficiency gain from combining data from multiple trials, compared to using data from each individual trial alone, through two simulation studies, one of which is designed based on a cholera vaccine trial previously carried out in Matlab, Bangladesh.
Assuntos
Vacinas contra Cólera , Cólera , Humanos , Ensaios Clínicos Controlados Aleatórios como Assunto , Cólera/prevenção & controle , Vacinação , Projetos de PesquisaRESUMO
The test-negative design (TND) is an observational study design to evaluate vaccine effectiveness (VE) that enrolls individuals receiving diagnostic testing for a target disease as part of routine care. VE is estimated as one minus the adjusted odds ratio of testing positive versus negative comparing vaccinated and unvaccinated patients. Although the TND is related to case-control studies, it is distinct in that the ratio of test-positive cases to test-negative controls is not typically pre-specified. For both types of studies, sparse cells are common when vaccines are highly effective. We consider the implications of these features on power for the TND. We use simulation studies to explore three hypothesis-testing procedures and associated sample size calculations for case-control and TND studies. These tests, all based on a simple logistic regression model, are a standard Wald test, a continuity-corrected Wald test, and a score test. The Wald test performs poorly in both case-control and TND when VE is high because the number of vaccinated test-positive cases can be low or zero. Continuity corrections help to stabilize the variance but induce bias. We observe superior performance with the score test as the variance is pooled under the null hypothesis of no group differences. We recommend using a score-based approach to design and analyze both case-control and TND. We propose a modification to the TND score sample size to account for additional variability in the ratio of controls over cases. This work enhances our understanding of the data generating mechanism in a test-negative design (TND) and how it is distinct from that of a case-control study due to its passive recruitment of controls.
Assuntos
Projetos de Pesquisa , Humanos , Tamanho da Amostra , Estudos de Casos e Controles , Eficácia de Vacinas/estatística & dados numéricos , Modelos Logísticos , Simulação por Computador , Razão de Chances , Vacinação/estatística & dados numéricos , Estudos Observacionais como Assunto/métodos , Estudos Observacionais como Assunto/estatística & dados numéricosRESUMO
BACKGROUND: The serial interval is the period of time between symptom onset in the primary case and symptom onset in the secondary case. Understanding the serial interval is important for determining transmission dynamics of infectious diseases like COVID-19, including the reproduction number and secondary attack rates, which could influence control measures. Early meta-analyses of COVID-19 reported serial intervals of 5.2 days (95% CI: 4.9-5.5) for the original wild-type variant and 5.2 days (95% CI: 4.87-5.47) for Alpha variant. The serial interval has been shown to decrease over the course of an epidemic for other respiratory diseases, which may be due to accumulating viral mutations and implementation of more effective nonpharmaceutical interventions. We therefore aggregated the literature to estimate serial intervals for Delta and Omicron variants. METHODS: This study followed Preferred Reporting Items for Systematic Reviews and Meta-analyses guidelines. A systematic literature search was conducted of PubMed, Scopus, Cochrane Library, ScienceDirect, and preprint server medRxiv for articles published from April 4, 2021, through May 23, 2023. Search terms were: ("serial interval" or "generation time"), ("Omicron" or "Delta"), and ("SARS-CoV-2" or "COVID-19"). Meta-analyses were done for Delta and Omicron variants using a restricted maximum-likelihood estimator model with a random effect for each study. Pooled average estimates and 95% confidence intervals (95% CI) are reported. RESULTS: There were 46,648 primary/secondary case pairs included for the meta-analysis of Delta and 18,324 for Omicron. Mean serial interval for included studies ranged from 2.3-5.8 days for Delta and 2.1-4.8 days for Omicron. The pooled mean serial interval for Delta was 3.9 days (95% CI: 3.4-4.3) (20 studies) and Omicron was 3.2 days (95% CI: 2.9-3.5) (20 studies). Mean estimated serial interval for BA.1 was 3.3 days (95% CI: 2.8-3.7) (11 studies), BA.2 was 2.9 days (95% CI: 2.7-3.1) (six studies), and BA.5 was 2.3 days (95% CI: 1.6-3.1) (three studies). CONCLUSIONS: Serial interval estimates for Delta and Omicron were shorter than ancestral SARS-CoV-2 variants. More recent Omicron subvariants had even shorter serial intervals suggesting serial intervals may be shortening over time. This suggests more rapid transmission from one generation of cases to the next, consistent with the observed faster growth dynamic of these variants compared to their ancestors. Additional changes to the serial interval may occur as SARS-CoV-2 continues to circulate and evolve. Changes to population immunity (due to infection and/or vaccination) may further modify it.
Assuntos
COVID-19 , Epidemias , Humanos , Família , SARS-CoV-2/genéticaRESUMO
BACKGROUND: An ongoing cluster-randomized trial for the prevention of arboviral diseases utilizes covariate-constrained randomization to balance two treatment arms across four specified covariates and geographic sector. Each cluster is within a census tract of the city of Mérida, Mexico, and there were 133 eligible tracts from which to select 50. As some selected clusters may have been subsequently found unsuitable in the field, we desired a strategy to substitute new clusters while maintaining covariate balance. METHODS: We developed an algorithm that successfully identified a subset of clusters that maximized the average minimum pairwise distance between clusters in order to reduce contamination and balanced the specified covariates both before and after substitutions were made. SIMULATIONS: Simulations were performed to explore some limitations of this algorithm. The number of selected clusters and eligible clusters were varied along with the method of selecting the final allocation pattern. CONCLUSION: The algorithm is presented here as a series of optional steps that can be added to the standard covariate-constrained randomization process in order to achieve spatial dispersion, cluster subsampling, and cluster substitution. Simulation results indicate that these extensions can be used without loss of statistical validity, given a sufficient number of clusters included in the trial.
Assuntos
Algoritmos , Projetos de Pesquisa , Humanos , Análise por Conglomerados , Distribuição Aleatória , Simulação por ComputadorRESUMO
BACKGROUND: The benefit of primary and booster vaccination in people who experienced a prior Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-CoV-2) infection remains unclear. The objective of this study was to estimate the effectiveness of primary (two-dose series) and booster (third dose) mRNA vaccination against Omicron (lineage BA.1) infection among people with a prior documented infection. METHODS AND FINDINGS: We conducted a test-negative case-control study of reverse transcription PCRs (RT-PCRs) analyzed with the TaqPath (Thermo Fisher Scientific) assay and recorded in the Yale New Haven Health system from November 1, 2021, to April 30, 2022. Overall, 11,307 cases (positive TaqPath analyzed RT-PCRs with S-gene target failure [SGTF]) and 130,041 controls (negative TaqPath analyzed RT-PCRs) were included (median age: cases: 35 years, controls: 39 years). Among cases and controls, 5.9% and 8.1% had a documented prior infection (positive SARS-CoV-2 test record ≥90 days prior to the included test), respectively. We estimated the effectiveness of primary and booster vaccination relative to SGTF-defined Omicron (lineage BA.1) variant infection using a logistic regression adjusted for date of test, age, sex, race/ethnicity, insurance, comorbidities, social venerability index, municipality, and healthcare utilization. The effectiveness of primary vaccination 14 to 149 days after the second dose was 41.0% (95% confidence interval (CI): 14.1% to 59.4%, p 0.006) and 27.1% (95% CI: 18.7% to 34.6%, p < 0.001) for people with and without a documented prior infection, respectively. The effectiveness of booster vaccination (≥14 days after booster dose) was 47.1% (95% CI: 22.4% to 63.9%, p 0.001) and 54.1% (95% CI: 49.2% to 58.4%, p < 0.001) in people with and without a documented prior infection, respectively. To test whether booster vaccination reduced the risk of infection beyond that of the primary series, we compared the odds of infection among boosted (≥14 days after booster dose) and booster-eligible people (≥150 days after second dose). The odds ratio (OR) comparing boosted and booster-eligible people with a documented prior infection was 0.79 (95% CI: 0.54 to 1.16, p 0.222), whereas the OR comparing boosted and booster-eligible people without a documented prior infection was 0.54 (95% CI: 0.49 to 0.59, p < 0.001). This study's limitations include the risk of residual confounding, the use of data from a single system, and the reliance on TaqPath analyzed RT-PCR results. CONCLUSIONS: In this study, we observed that primary vaccination provided significant but limited protection against Omicron (lineage BA.1) infection among people with and without a documented prior infection. While booster vaccination was associated with additional protection against Omicron BA.1 infection in people without a documented prior infection, it was not found to be associated with additional protection among people with a documented prior infection. These findings support primary vaccination in people regardless of documented prior infection status but suggest that infection history may impact the relative benefit of booster doses.
Assuntos
COVID-19 , Humanos , Adulto , COVID-19/epidemiologia , COVID-19/prevenção & controle , SARS-CoV-2/genética , Estudos de Casos e Controles , Razão de Chances , VacinaçãoRESUMO
Postauthorization observational studies play a key role in understanding COVID-19 vaccine effectiveness following the demonstration of efficacy in clinical trials. Although bias due to confounding, selection bias, and misclassification can be mitigated through careful study design, unmeasured confounding is likely to remain in these observational studies. Phase III trials of COVID-19 vaccines have shown that protection from vaccination does not occur immediately, meaning that COVID-19 risk should be similar in recently vaccinated and unvaccinated individuals, in the absence of confounding or other bias. Several studies have used the estimated effectiveness among recently vaccinated individuals as a negative control exposure to detect bias in vaccine effectiveness estimates. In this paper, we introduce a theoretical framework to describe the interpretation of such a bias indicator in test-negative studies, and outline strong assumptions that would allow vaccine effectiveness among recently vaccinated individuals to serve as a negative control exposure.
Assuntos
Vacinas contra COVID-19 , COVID-19 , COVID-19/prevenção & controle , Vacinas contra COVID-19/uso terapêutico , Estudos de Casos e Controles , Humanos , Vacinação , Eficácia de VacinasRESUMO
The ring vaccination trial is a recently developed approach for evaluating the efficacy and effectiveness of vaccines, modeled after the surveillance and containment strategy of ring vaccination. Contacts and contacts of contacts of a newly identified disease case form a ring, and these rings are randomized as part of a cluster-randomized trial or with individual randomization within rings. Key advantages of the design include its flexibility to follow the epidemic as it progresses and the targeting of high-risk participants to increase power. We describe the application of the design to estimate the efficacy and effectiveness of an Ebola vaccine during the 2014-2016 West African Ebola epidemic. The design has several notable statistical features. Because vaccination occurs around the time of exposure, the design is particularly sensitive to the choice of per protocol analysis period. If incidence wanes before the per protocol analysis period begins (due to a slow-acting vaccine or a fast-moving pathogen), power can be substantially reduced. Mathematical modeling is valuable for exploring the suitability of the approach in different disease settings. Another statistical feature is zero inflation, which can occur if the chain of transmission does not take off within a ring. In the application to Ebola, the majority of rings had zero subsequent cases. The ring vaccination trial can be extended in several ways, including the definition of rings (e.g. contact-based, spatial, and occupational). The design will be valuable in settings where the spatio-temporal spread of the pathogen is highly focused and unpredictable.
Assuntos
Vacinas contra Ebola , Doença pelo Vírus Ebola , Surtos de Doenças/prevenção & controle , Vacinas contra Ebola/uso terapêutico , Doença pelo Vírus Ebola/epidemiologia , Doença pelo Vírus Ebola/prevenção & controle , Humanos , Vacinação/métodos , Eficácia de VacinasRESUMO
BACKGROUND: Previous studies have demonstrated that decreased impingement-free range of motion (ROM) can adversely influence clinical outcomes following reverse shoulder arthroplasty (RSA). Inferior placement of the glenosphere is thought to minimize impingement and its associated sequelae. This study evaluated the relationship between inferior overhang of the glenosphere and clinical outcomes in patients undergoing primary RSA using a lateralized humeral implant design. METHODS: By use of a prospectively collected shoulder arthroplasty database, all primary RSAs performed at our institution between 2007 and 2015 with a single implant design (lateralized humerus and medialized glenoid) and minimum 2-year follow-up were evaluated. Glenosphere overhang in relation to the inferior rim of the glenoid was measured in millimeters on postoperative Grashey radiographs of the shoulder and categorized into tertiles (low, <7.1 mm; medium, 7.1 to 9.9 mm; and high, >9.9 mm). Clinical outcomes of interest comprised the changes between preoperative and postoperative values in the following ROM and outcome score measures: active forward elevation (aFE), active external rotation, American Shoulder and Elbow Surgeons score, Constant-Murley score, Shoulder Pain and Disability Index score, and Simple Shoulder Test score. Random-effects linear models were used to assess univariate and multivariable associations between overhang tertile and change in patient outcomes. Differences in outcomes were further compared using the minimal clinically important difference (MCID). RESULTS: The study identified 284 shoulders in 265 patients. The median follow-up period was 36 months (range, 24-108 months). The median glenosphere inferior overhang was 8.4 mm, with an interquartile range of 6.3-10.6 mm. Plots demonstrated nonlinear relationships between overhang and outcome scores and between overhang and ROM. Patients with high overhang experienced a significantly greater improvement in aFE compared with patients with low overhang (P = .019), which exceeded the MCID. No other differences in ROM and outcome scores between overhang groups exceeded the MCID. For other outcome scores and ROM measurements, there was no significant relationship with glenosphere overhang. Increased overhang was associated with a significantly lower incidence of scapular notching (P = .005). CONCLUSION: Patients undergoing RSA using a lateralized humerus design with greater inferior overhang of the glenosphere demonstrated a significantly greater improvement in aFE and lower rate of notching compared with those with low overhang. No ideal glenosphere overhang range was identified to maximize function in this study.
Assuntos
Artroplastia do Ombro , Articulação do Ombro , Prótese de Ombro , Artroplastia do Ombro/efeitos adversos , Humanos , Úmero/diagnóstico por imagem , Úmero/cirurgia , Desenho de Prótese , Amplitude de Movimento Articular , Articulação do Ombro/diagnóstico por imagem , Articulação do Ombro/cirurgia , Prótese de Ombro/efeitos adversos , Resultado do TratamentoRESUMO
Comparison of coronavirus disease 2019 (COVID-19) case numbers over time and between locations is complicated by limits to virological testing to confirm severe acute respiratory syndrome coronavirus 2 infection. The proportion of tested individuals who have tested positive (test-positive proportion, TPP) can potentially be used to inform trends in incidence. We propose a model for testing in a population experiencing an epidemic of COVID-19 and derive an expression for TPP in terms of well-defined parameters related to testing and presence of other pathogens causing COVID-19-like symptoms. In the absence of dramatic shifts of testing practices in time or between locations, the TPP is positively correlated with the incidence of infection. We show that the proportion of tested individuals who present COVID-19-like symptoms encodes information similar to the TPP but has different relationships with the testing parameters, and can thus provide additional information regarding dynamic changes in TPP and incidence. Finally, we compare data on confirmed cases and TPP from US states up to October 2020. We conjecture why states might have higher or lower TPP than average. Collection of symptom status and age/risk category of tested individuals can increase the utility of TPP in assessing the state of the pandemic in different locations and times.
Assuntos
Teste para COVID-19 , COVID-19/diagnóstico , COVID-19/epidemiologia , COVID-19/transmissão , Modelos Teóricos , Vigilância da População/métodos , Humanos , Incidência , Pandemias , SARS-CoV-2RESUMO
Observational studies of the effectiveness of vaccines to prevent COVID-19 are needed to inform real-world use. Such studies are now underway amid the ongoing rollout of SARS-CoV-2 vaccines globally. Although traditional case-control and test-negative design studies feature prominently among strategies used to assess vaccine effectiveness, such studies may encounter important threats to validity. Here, we review the theoretical basis for estimation of vaccine direct effects under traditional case-control and test-negative design frameworks, addressing specific natural history parameters of SARS-CoV-2 infection and COVID-19 relevant to these designs. Bias may be introduced by misclassification of cases and controls, particularly when clinical case criteria include common, nonspecific indicators of COVID-19. When using diagnostic assays with high analytical sensitivity for SARS-CoV-2 detection, individuals testing positive may be counted as cases even if their symptoms are due to other causes. The traditional case-control design may be particularly prone to confounding due to associations of vaccination with healthcare-seeking behavior or risk of infection. The test-negative design reduces but may not eliminate this confounding, for instance, if individuals who receive vaccination seek care or testing for less-severe illness. These circumstances indicate the two study designs cannot be applied naively to datasets gathered through public health surveillance or administrative sources. We suggest practical strategies to reduce bias in vaccine effectiveness estimates at the study design and analysis stages.
Assuntos
COVID-19 , Vacinas , Vacinas contra COVID-19 , Humanos , Estudos Retrospectivos , SARS-CoV-2RESUMO
BACKGROUND: Novel strategies are needed to make vaccine efficacy trials more robust given uncertain epidemiology of infectious disease outbreaks, such as arboviruses like Zika. Spatially resolved mathematical and statistical models can help investigators identify sites at highest risk of future transmission and prioritize these for inclusion in trials. Models can also characterize uncertainty in whether transmission will occur at a site, and how nearby or connected sites may have correlated outcomes. A structure is needed for how trials can use models to address key design questions, including how to prioritize sites, the optimal number of sites, and how to allocate participants across sites. METHODS: We illustrate the added value of models using the motivating example of Zika vaccine trial planning during the 2015-2017 Zika epidemic. We used a stochastic, spatially resolved, transmission model (the Global Epidemic and Mobility model) to simulate epidemics and site-level incidence at 100 high-risk sites in the Americas. We considered several strategies for prioritizing sites (average site-level incidence of infection across epidemics, median incidence, probability of exceeding 1% incidence), selecting the number of sites, and allocating sample size across sites (equal enrollment, proportional to average incidence, proportional to rank). To evaluate each design, we stochastically simulated trials in each hypothetical epidemic by drawing observed cases from site-level incidence data. RESULTS: When constraining overall trial size, the optimal number of sites represents a balance between prioritizing highest-risk sites and having enough sites to reduce the chance of observing too few endpoints. The optimal number of sites remained roughly constant regardless of the targeted number of events, although it is necessary to increase the sample size to achieve the desired power. Though different ranking strategies returned different site orders, they performed similarly with respect to trial power. Instead of enrolling participants equally from each site, investigators can allocate participants proportional to projected incidence, though this did not provide an advantage in our example because the top sites had similar risk profiles. Sites from the same geographic region may have similar outcomes, so optimal combinations of sites may be geographically dispersed, even when these are not the highest ranked sites. CONCLUSION: Mathematical and statistical models may assist in designing successful vaccination trials by capturing uncertainty and correlation in future transmission. Although many factors affect site selection, such as logistical feasibility, models can help investigators optimize site selection and the number and size of participating sites. Although our study focused on trial design for an emerging arbovirus, a similar approach can be made for any infectious disease with the appropriate model for the particular disease.
Assuntos
Epidemias , Vacinas , Infecção por Zika virus , Zika virus , Humanos , Incidência , Modelos Estatísticos , Tamanho da Amostra , Infecção por Zika virus/epidemiologia , Infecção por Zika virus/prevenção & controleRESUMO
In the test-negative design, routine testing at health-care facilities is leveraged to estimate the effectiveness of an intervention such as a vaccine. The odds of vaccination for individuals who test positive for a target pathogen is compared with the odds of vaccination for individuals who test negative for that pathogen, adjusting for key confounders. The design is rapidly growing in popularity, but many open questions remain about its properties. In this paper, we examine temporal confounding by generalizing derivations to allow for time-varying vaccine status, including out-of-season controls, and open populations. We confirm that calendar time is an important confounder when vaccine status varies during the study. We demonstrate that, where time is not a confounder, including out-of-season controls can improve precision. We generalize these results to open populations. We use our theoretical findings to interpret 3 recent papers utilizing the test-negative design. Through careful examination of the theoretical properties of this study design, we provide key insights that can directly inform the implementation and analysis of future test-negative studies.
Assuntos
Controle de Doenças Transmissíveis/métodos , Doenças Transmissíveis/epidemiologia , Infectologia/métodos , Fatores de Tempo , Vacinação/estatística & dados numéricos , Fatores de Confusão Epidemiológicos , Humanos , Modelos Teóricos , Projetos de Pesquisa , Estações do AnoRESUMO
We propose an adaptive enrichment approach to test an active factor, which is a factor whose effect is non-zero in at least one subpopulation. We implement a two-stage play-the-winner design where all subjects in the second stage are enrolled from the subpopulation that has the highest observed effect in the first stage. We recommend a weighted Fisher's combination of the most powerful test for each stage, respectively: the first stage Hotelling's test and the second stage noncentral chi-square test. The test is further extended to cover binary outcomes and time-to-event outcomes.
Assuntos
Ensaios Clínicos Adaptados como Assunto/estatística & dados numéricos , Projetos de Pesquisa/estatística & dados numéricos , Catastrofização/genética , Catastrofização/psicologia , Catecol O-Metiltransferase/genética , Interpretação Estatística de Dados , Humanos , Modelos Estatísticos , Polimorfismo de Nucleotídeo Único , Dor de Ombro/genética , Dor de Ombro/psicologiaRESUMO
We use a data-driven global stochastic epidemic model to analyze the spread of the Zika virus (ZIKV) in the Americas. The model has high spatial and temporal resolution and integrates real-world demographic, human mobility, socioeconomic, temperature, and vector density data. We estimate that the first introduction of ZIKV to Brazil likely occurred between August 2013 and April 2014 (90% credible interval). We provide simulated epidemic profiles of incident ZIKV infections for several countries in the Americas through February 2017. The ZIKV epidemic is characterized by slow growth and high spatial and seasonal heterogeneity, attributable to the dynamics of the mosquito vector and to the characteristics and mobility of the human populations. We project the expected timing and number of pregnancies infected with ZIKV during the first trimester and provide estimates of microcephaly cases assuming different levels of risk as reported in empirical retrospective studies. Our approach represents a modeling effort aimed at understanding the potential magnitude and timing of the ZIKV epidemic and it can be potentially used as a template for the analysis of future mosquito-borne epidemics.
Assuntos
Infecção por Zika virus/epidemiologia , Aedes/virologia , América/epidemiologia , Animais , Brasil/epidemiologia , Epidemias , Feminino , Humanos , Recém-Nascido , Masculino , Microcefalia/complicações , Microcefalia/epidemiologia , Modelos Biológicos , Modelos Estatísticos , Mosquitos Vetores/virologia , Gravidez , Complicações Infecciosas na Gravidez/epidemiologia , Estudos Retrospectivos , Fatores de Risco , Processos Estocásticos , Zika virus/isolamento & purificação , Infecção por Zika virus/transmissãoAssuntos
Vacinas contra COVID-19 , COVID-19 , Estudos de Casos e Controles , Hospitalização , Humanos , SARS-CoV-2RESUMO
Sierra Leone is the most severely affected country by an unprecedented outbreak of Ebola virus disease (EVD) in West Africa. Although successfully contained, the transmission dynamics of EVD and the impact of interventions in the country remain unclear. We established a database of confirmed and suspected EVD cases from May 2014 to September 2015 in Sierra Leone and mapped the spatiotemporal distribution of cases at the chiefdom level. A Poisson transmission model revealed that the transmissibility at the chiefdom level, estimated as the average number of secondary infections caused by a patient per week, was reduced by 43% [95% confidence interval (CI): 30%, 52%] after October 2014, when the strategic plan of the United Nations Mission for Emergency Ebola Response was initiated, and by 65% (95% CI: 57%, 71%) after the end of December 2014, when 100% case isolation and safe burials were essentially achieved, both compared with before October 2014. Population density, proximity to Ebola treatment centers, cropland coverage, and atmospheric temperature were associated with EVD transmission. The household secondary attack rate (SAR) was estimated to be 0.059 (95% CI: 0.050, 0.070) for the overall outbreak. The household SAR was reduced by 82%, from 0.093 to 0.017, after the nationwide campaign to achieve 100% case isolation and safe burials had been conducted. This study provides a complete overview of the transmission dynamics of the 2014-2015 EVD outbreak in Sierra Leone at both chiefdom and household levels. The interventions implemented in Sierra Leone seem effective in containing the epidemic, particularly in interrupting household transmission.
Assuntos
Bases de Dados Factuais , Ebolavirus , Doença pelo Vírus Ebola/epidemiologia , Doença pelo Vírus Ebola/terapia , Doença pelo Vírus Ebola/transmissão , Modelos Biológicos , Feminino , Humanos , Masculino , Serra Leoa/epidemiologiaRESUMO
We conducted a 3-year longitudinal serologic survey on an open cohort of poultry workers, swine workers, and general population controls to assess avian influenza A virus (AIV) seroprevalence and seroincidence and virologic diversity at live poultry markets (LPMs) in Wuxi City, Jiangsu Province, China. Of 964 poultry workers, 9 (0.93%) were seropositive for subtype H7N9 virus, 18 (1.87%) for H9N2, and 18 (1.87%) for H5N1. Of 468 poultry workers followed longitudinally, 2 (0.43%), 13 (2.78%), and 7 (1.5%) seroconverted, respectively; incidence was 1.27, 8.28, and 4.46/1,000 person-years for H7N9, H9N2, and H5N1 viruses, respectively. Longitudinal surveillance of AIVs at 9 LPMs revealed high co-circulation of H9, H7, and H5 subtypes. We detected AIVs in 726 (23.3%) of 3,121 samples and identified a high diversity (10 subtypes) of new genetic constellations and reassortant viruses. These data suggest that stronger surveillance for AIVs within LPMs and high-risk populations is imperative.