Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 89
Filtrar
1.
BMC Biol ; 21(1): 189, 2023 09 07.
Artículo en Inglés | MEDLINE | ID: mdl-37674179

RESUMEN

BACKGROUND: Researchers performing high-quality systematic reviews search across multiple databases to identify relevant evidence. However, the same publication is often retrieved from several databases. Identifying and removing such duplicates ("deduplication") can be extremely time-consuming, but failure to remove these citations can lead to the wrongful inclusion of duplicate data. Many existing tools are not sensitive enough, lack interoperability with other tools, are not freely accessible, or are difficult to use without programming knowledge. Here, we report the performance of our Automated Systematic Search Deduplicator (ASySD), a novel tool to perform automated deduplication of systematic searches for biomedical reviews. METHODS: We evaluated ASySD's performance on 5 unseen biomedical systematic search datasets of various sizes (1845-79,880 citations). We compared the performance of ASySD with EndNote's automated deduplication option and with the Systematic Review Assistant Deduplication Module (SRA-DM). RESULTS: ASySD identified more duplicates than either SRA-DM or EndNote, with a sensitivity in different datasets of 0.95 to 0.99. The false-positive rate was comparable to human performance, with a specificity of > 0.99. The tool took less than 1 h to identify and remove duplicates within each dataset. CONCLUSIONS: For duplicate removal in biomedical systematic reviews, ASySD is a highly sensitive, reliable, and time-saving tool. It is open source and freely available online as both an R package and a user-friendly web application.


Asunto(s)
Programas Informáticos , Revisiones Sistemáticas como Asunto , Humanos , Proyectos de Investigación
2.
Eur Stroke J ; 8(4): 923-931, 2023 Dec.
Artículo en Inglés | MEDLINE | ID: mdl-37606096

RESUMEN

INTRODUCTION: The prevalence of outcome reporting bias (ORB, i.e. selective reporting according to the results observed) across primary outcomes in randomised controlled trials (RCTs) including participants with stroke or transient ischaemic attack (TIA) is unknown. MATERIALS AND METHODS: We searched the Cochrane Database of Systematic Reviews on 3 February 2021 for reviews published 2008-2020 with at least one RCT of a therapeutic intervention, for participants with stroke or TIA, and a safety or efficacy outcome. We took a random sample of these RCTs and included those with a trial registry record or protocol published before reporting results. Two reviewers assessed discrepancies in outcome reporting across the trial registry record, protocol, statistical analysis plan, and publication for each RCT, using the classification system designed by the Outcome Reporting Bias in Trials group. RESULTS: Of 600 RCTs, we identified a trial registry record in 120 (20%), a protocol in 28 (5%), and a statistical analysis plan in 5 (1%) with 123 (21%) distinct RCTs being eligible for assessment: 110 (89%, 95% CI 83-94) were at no risk, 7 (6%, 95% CI 3-11) RCTs were at low risk, and 6 (5%, 95% CI 2-10) were at high risk of ORB. DISCUSSION: The prevalence of ORB in primary outcomes was low in stroke/TIA RCTs that were included in Cochrane reviews and had an identifiable trial registry record or protocol. Concerningly, we were unable to identify a trial registry record or protocol in most of our sample. CONCLUSION: Work is needed to further reduce ORB in stroke/TIA RCTs and explore the generalisability of these findings to RCTs outside of Cochrane reviews or without a registry record or protocol, as well as to secondary outcomes.


Asunto(s)
Ataque Isquémico Transitorio , Accidente Cerebrovascular , Humanos , Sesgo , Ataque Isquémico Transitorio/epidemiología , Ensayos Clínicos Controlados Aleatorios como Asunto , Accidente Cerebrovascular/epidemiología
3.
J Endocrinol ; 258(1)2023 07 01.
Artículo en Inglés | MEDLINE | ID: mdl-37074416

RESUMEN

In biomedicine and many other fields, there are growing concerns around the reproducibility of research findings, with many researchers being unable to replicate their own or others' results. This raises important questions as to the validity and usefulness of much published research. In this review, we aim to engage researchers in the issue of research reproducibility and equip them with the necessary tools to increase the reproducibility of their research. We first highlight the causes and potential impact of non-reproducible research and emphasise the benefits of working reproducibly for the researcher and broader research community. We address specific targets for improvement and steps that individual researchers can take to increase the reproducibility of their work. We next provide recommendations for improving the design and conduct of experiments, focusing on in vivo animal experiments. We describe common sources of poor internal validity of experiments and offer practical guidance for limiting these potential sources of bias at different experimental stages, as well as discussing other important considerations during experimental design. We provide a list of key resources available to researchers to improve experimental design, conduct, and reporting. We then discuss the importance of open research practices such as study preregistration and the use of preprints and describe recommendations around data management and sharing. Our review emphasises the importance of reproducible work and aims to empower every individual researcher to contribute to the reproducibility of research in their field.


Asunto(s)
Experimentación Animal , Animales , Reproducibilidad de los Resultados , Proyectos de Investigación
4.
Clin Sci (Lond) ; 137(2): 181-193, 2023 01 31.
Artículo en Inglés | MEDLINE | ID: mdl-36630537

RESUMEN

OBJECTIVE: Existing strategies to identify relevant studies for systematic review may not perform equally well across research domains. We compare four approaches based on either human or automated screening of either title and abstract or full text, and report the training of a machine learning algorithm to identify in vitro studies from bibliographic records. METHODS: We used a systematic review of oxygen-glucose deprivation (OGD) in PC-12 cells to compare approaches. For human screening, two reviewers independently screened studies based on title and abstract or full text, with disagreements reconciled by a third. For automated screening, we applied text mining to either title and abstract or full text. We trained a machine learning algorithm with decisions from 2000 randomly selected PubMed Central records enriched with a dataset of known in vitro studies. RESULTS: Full-text approaches performed best, with human (sensitivity: 0.990, specificity: 1.000 and precision: 0.994) outperforming text mining (sensitivity: 0.972, specificity: 0.980 and precision: 0.764). For title and abstract, text mining (sensitivity: 0.890, specificity: 0.995 and precision: 0.922) outperformed human screening (sensitivity: 0.862, specificity: 0.998 and precision: 0.975). At our target sensitivity of 95% the algorithm performed with specificity of 0.850 and precision of 0.700. CONCLUSION: In this in vitro systematic review, human screening based on title and abstract erroneously excluded 14% of relevant studies, perhaps because title and abstract provide an incomplete description of methods used. Our algorithm might be used as a first selection phase in in vitro systematic reviews to limit the extent of full text screening required.


Asunto(s)
Algoritmos , Minería de Datos , Humanos , Minería de Datos/métodos , Proyectos de Investigación , Aprendizaje Automático , Glucosa
5.
PLoS Biol ; 21(1): e3001949, 2023 01.
Artículo en Inglés | MEDLINE | ID: mdl-36693044

RESUMEN

The state of open science needs to be monitored to track changes over time and identify areas to create interventions to drive improvements. In order to monitor open science practices, they first need to be well defined and operationalized. To reach consensus on what open science practices to monitor at biomedical research institutions, we conducted a modified 3-round Delphi study. Participants were research administrators, researchers, specialists in dedicated open science roles, and librarians. In rounds 1 and 2, participants completed an online survey evaluating a set of potential open science practices, and for round 3, we hosted two half-day virtual meetings to discuss and vote on items that had not reached consensus. Ultimately, participants reached consensus on 19 open science practices. This core set of open science practices will form the foundation for institutional dashboards and may also be of value for the development of policy, education, and interventions.


Asunto(s)
Investigación Biomédica , Humanos , Consenso , Técnica Delphi , Encuestas y Cuestionarios , Proyectos de Investigación
6.
Wellcome Open Res ; 8: 365, 2023.
Artículo en Inglés | MEDLINE | ID: mdl-38634067

RESUMEN

BACKGROUND: There is an urgent need to develop more effective and safer antipsychotics beyond dopamine 2 receptor antagonists. An emerging and promising approach is TAAR1 agonism. Therefore, we will conduct a living systematic review and meta-analysis to synthesize and triangulate the evidence from preclinical animal experiments and clinical studies on the efficacy, safety, and underlying mechanism of action of TAAR1 agonism for psychosis. METHODS: Independent searches will be conducted in multiple electronic databases to identify clinical and animal experimental studies comparing TAAR1 agonists with licensed antipsychotics or other control conditions in individuals with psychosis or animal models for psychosis, respectively. The primary outcomes will be overall psychotic symptoms and their behavioural proxies in animals. Secondary outcomes will include side effects and neurobiological measures. Two independent reviewers will conduct study selection, data extraction using predefined forms, and risk of bias assessment using suitable tools based on the study design. Ontologies will be developed to facilitate study identification and data extraction. Data from clinical and animal studies will be synthesized separately using random-effects meta-analysis if appropriate, or synthesis without meta-analysis. Study characteristics will be investigated as potential sources of heterogeneity. Confidence in the evidence for each outcome and source of evidence will be evaluated, considering the summary of the association, potential concerns regarding internal and external validity, and reporting biases. When multiple sources of evidence are available for an outcome, an overall conclusion will be drawn in a triangulation meeting involving a multidisciplinary team of experts. We plan trimonthly updates of the review, and any modifications in the protocol will be documented. The review will be co-produced by multiple stakeholders aiming to produce impactful and relevant results and bridge the gap between preclinical and clinical research on psychosis. PROTOCOL REGISTRATION: PROSPERO-ID: CRD42023451628.

7.
Pain ; 163(11): 2076-2102, 2022 11 01.
Artículo en Inglés | MEDLINE | ID: mdl-35353780

RESUMEN

ABSTRACT: Burrowing behaviour is used to assess pain-associated behaviour in laboratory rodents. To gain insight into how models of disease-associated persistent pain and analgesics affect burrowing behaviour, we performed a systematic review and meta-analysis of studies that assessed burrowing behaviour. A systematic search in March 2020 and update in September 2020 was conducted in 4 databases. Study design characteristics and experimental data were extracted, followed by a random-effects meta-analysis. We explored the association between burrowing and monofilament-induced limb withdrawal. Dose response relationship was investigated for some analgesics. Forty-five studies were included in the meta-analysis, in which 16 model types and 14 drug classes were used. Most experiments used rat (79%) and male (72%) animals. Somatic inflammation and trauma-induced neuropathy models were associated with reduced burrowing behaviour. Analgesics (nonsteroidal anti-inflammatory drug and gabapentinoids) attenuated burrowing deficits in these models. Reporting of measures to reduce risk of bias was unclear except for randomisation which was high. There was not a correlation ( R2 = 0.1421) between burrowing and monofilament-induced limb withdrawal. Opioids, gabapentin, and naproxen showed reduced burrowing behaviour at high doses, whereas ibuprofen and celecoxib showed opposite trend. The findings indicate that burrowing could be used to assess pain-associated behaviour. We support the use of a portfolio of composite measures including spontaneous and stimulus-evoked tests. The information collected here could help in designing experiments involving burrowing assessment in models of disease-associated pain.


Asunto(s)
Ibuprofeno , Naproxeno , Analgésicos/farmacología , Analgésicos/uso terapéutico , Animales , Antiinflamatorios no Esteroideos/uso terapéutico , Conducta Animal/fisiología , Celecoxib/uso terapéutico , Modelos Animales de Enfermedad , Gabapentina/farmacología , Gabapentina/uso terapéutico , Ibuprofeno/farmacología , Ibuprofeno/uso terapéutico , Masculino , Naproxeno/uso terapéutico , Dolor/tratamiento farmacológico , Dolor/etiología , Ratas , Roedores
8.
Syst Rev ; 10(1): 292, 2021 11 02.
Artículo en Inglés | MEDLINE | ID: mdl-34727973

RESUMEN

BACKGROUND: Time-to-event data is frequently reported in both clinical and preclinical research spheres. Systematic review and meta-analysis is a tool that can help to identify pitfalls in preclinical research conduct and reporting that can help to improve translational efficacy. However, pooling of studies using hazard ratios (HRs) is cumbersome especially in preclinical meta-analyses including large numbers of small studies. Median survival is a much simpler metric although because of some limitations, which may not apply to preclinical data, it is generally not used in survival meta-analysis. We aimed to appraise its performance when compared with hazard ratio-based meta-analysis when pooling large numbers of small, imprecise studies. METHODS: We simulated a survival dataset with features representative of a typical preclinical survival meta-analysis, including with influence of a treatment and a number of covariates. We calculated individual patient data-based hazard ratios and median survival ratios (MSRs), comparing the summary statistics directly and their performance at random-effects meta-analysis. Finally, we compared their sensitivity to detect associations between treatment and influential covariates at meta-regression. RESULTS: There was an imperfect correlation between MSR and HR, although the opposing direction of treatment effects between summary statistics appeared not to be a major issue. Precision was more conservative for HR than MSR, meaning that estimates of heterogeneity were lower. There was a slight sensitivity advantage for MSR at meta-analysis and meta-regression, although power was low in all circumstances. CONCLUSIONS: We believe we have validated MSR as a summary statistic for use in a meta-analysis of small, imprecise experimental survival studies-helping to increase confidence and efficiency in future reviews in this area. While assessment of study precision and therefore weighting is less reliable, MSR appears to perform favourably during meta-analysis. Sensitivity of meta-regression was low for this set of parameters, so pooling of treatments to increase sample size may be required to ensure confidence in preclinical survival meta-regressions.


Asunto(s)
Análisis de Supervivencia , Humanos
9.
PLoS Biol ; 19(5): e3001177, 2021 05.
Artículo en Inglés | MEDLINE | ID: mdl-33951050

RESUMEN

In an effort to better utilize published evidence obtained from animal experiments, systematic reviews of preclinical studies are increasingly more common-along with the methods and tools to appraise them (e.g., SYstematic Review Center for Laboratory animal Experimentation [SYRCLE's] risk of bias tool). We performed a cross-sectional study of a sample of recent preclinical systematic reviews (2015-2018) and examined a range of epidemiological characteristics and used a 46-item checklist to assess reporting details. We identified 442 reviews published across 43 countries in 23 different disease domains that used 26 animal species. Reporting of key details to ensure transparency and reproducibility was inconsistent across reviews and within article sections. Items were most completely reported in the title, introduction, and results sections of the reviews, while least reported in the methods and discussion sections. Less than half of reviews reported that a risk of bias assessment for internal and external validity was undertaken, and none reported methods for evaluating construct validity. Our results demonstrate that a considerable number of preclinical systematic reviews investigating diverse topics have been conducted; however, their quality of reporting is inconsistent. Our study provides the justification and evidence to inform the development of guidelines for conducting and reporting preclinical systematic reviews.


Asunto(s)
Revisión de la Investigación por Pares/métodos , Revisión de la Investigación por Pares/normas , Proyectos de Investigación/normas , Experimentación Animal/normas , Animales , Sesgo , Lista de Verificación/normas , Evaluación Preclínica de Medicamentos/métodos , Evaluación Preclínica de Medicamentos/normas , Investigación Empírica , Métodos Epidemiológicos , Epidemiología/tendencias , Humanos , Revisión de la Investigación por Pares/tendencias , Publicaciones , Reproducibilidad de los Resultados , Proyectos de Investigación/tendencias
10.
BMJ Open Sci ; 5(1): e100103, 2021.
Artículo en Inglés | MEDLINE | ID: mdl-35047698

RESUMEN

Preclinical research is a vital step in the drug discovery pipeline and more generally in helping to better understand human disease aetiology and its management. Systematic reviews (SRs) can be powerful in summarising and appraising this evidence concerning a specific research question, to highlight areas of improvements, areas for further research and areas where evidence may be sufficient to take forward to other research domains, for instance clinical trial. Guidance and tools for preclinical research synthesis remain limited despite their clear utility. We aimed to create an online end-to-end platform primarily for conducting SRs of preclinical studies, that was flexible enough to support a wide variety of experimental designs, was adaptable to different research questions, would allow users to adopt emerging automated tools and support them during their review process using best practice. In this article, we introduce the Systematic Review Facility (https://syrf.org.uk), which was launched in 2016 and designed to support primarily preclinical SRs from small independent projects to large, crowdsourced projects. We discuss the architecture of the app and its features, including the opportunity to collaborate easily, to efficiently manage projects, to screen and annotate studies for important features (metadata), to extract outcome data into a secure database, and tailor these steps to each project. We introduce how we are working to leverage the use of automation tools and allow the integration of these services to accelerate and automate steps in the systematic review workflow.

11.
BMJ Open Sci ; 5(1): e100135, 2021.
Artículo en Inglés | MEDLINE | ID: mdl-35047702

RESUMEN

OBJECTIVE: Thigmotaxis is an innate predator avoidance behaviour of rodents and is enhanced when animals are under stress. It is characterised by the preference of a rodent to seek shelter, rather than expose itself to the aversive open area. The behaviour has been proposed to be a measurable construct that can address the impact of pain on rodent behaviour. This systematic review will assess whether thigmotaxis can be influenced by experimental persistent pain and attenuated by pharmacological interventions in rodents. SEARCH STRATEGY: We will conduct search on three electronic databases to identify studies in which thigmotaxis was used as an outcome measure contextualised to a rodent model associated with persistent pain. All studies published until the date of the search will be considered. SCREENING AND ANNOTATION: Two independent reviewers will screen studies based on the order of (1) titles and abstracts, and (2) full texts. DATA MANAGEMENT AND REPORTING: For meta-analysis, we will extract thigmotactic behavioural data and calculate effect sizes. Effect sizes will be combined using a random-effects model. We will assess heterogeneity and identify sources of heterogeneity. A risk-of-bias assessment will be conducted to evaluate study quality. Publication bias will be assessed using funnel plots, Egger's regression and trim-and-fill analysis. We will also extract stimulus-evoked limb withdrawal data to assess its correlation with thigmotaxis in the same animals. The evidence obtained will provide a comprehensive understanding of the strengths and limitations of using thigmotactic outcome measure in animal pain research so that future experimental designs can be optimised. We will follow the Preferred Reporting Items for Systematic Reviews and Meta-Analyses reporting guidelines and disseminate the review findings through publication and conference presentation.

12.
PLoS Biol ; 18(7): e3000410, 2020 07.
Artículo en Inglés | MEDLINE | ID: mdl-32663219

RESUMEN

Reproducible science requires transparent reporting. The ARRIVE guidelines (Animal Research: Reporting of In Vivo Experiments) were originally developed in 2010 to improve the reporting of animal research. They consist of a checklist of information to include in publications describing in vivo experiments to enable others to scrutinise the work adequately, evaluate its methodological rigour, and reproduce the methods and results. Despite considerable levels of endorsement by funders and journals over the years, adherence to the guidelines has been inconsistent, and the anticipated improvements in the quality of reporting in animal research publications have not been achieved. Here, we introduce ARRIVE 2.0. The guidelines have been updated and information reorganised to facilitate their use in practice. We used a Delphi exercise to prioritise and divide the items of the guidelines into 2 sets, the "ARRIVE Essential 10," which constitutes the minimum requirement, and the "Recommended Set," which describes the research context. This division facilitates improved reporting of animal research by supporting a stepwise approach to implementation. This helps journal editors and reviewers verify that the most important items are being reported in manuscripts. We have also developed the accompanying Explanation and Elaboration (E&E) document, which serves (1) to explain the rationale behind each item in the guidelines, (2) to clarify key concepts, and (3) to provide illustrative examples. We aim, through these changes, to help ensure that researchers, reviewers, and journal editors are better equipped to improve the rigour and transparency of the scientific process and thus reproducibility.


Asunto(s)
Experimentación Animal , Guías como Asunto , Informe de Investigación , Animales , Lista de Verificación
13.
PLoS Biol ; 18(7): e3000411, 2020 07.
Artículo en Inglés | MEDLINE | ID: mdl-32663221

RESUMEN

Improving the reproducibility of biomedical research is a major challenge. Transparent and accurate reporting is vital to this process; it allows readers to assess the reliability of the findings and repeat or build upon the work of other researchers. The ARRIVE guidelines (Animal Research: Reporting In Vivo Experiments) were developed in 2010 to help authors and journals identify the minimum information necessary to report in publications describing in vivo experiments. Despite widespread endorsement by the scientific community, the impact of ARRIVE on the transparency of reporting in animal research publications has been limited. We have revised the ARRIVE guidelines to update them and facilitate their use in practice. The revised guidelines are published alongside this paper. This explanation and elaboration document was developed as part of the revision. It provides further information about each of the 21 items in ARRIVE 2.0, including the rationale and supporting evidence for their inclusion in the guidelines, elaboration of details to report, and examples of good reporting from the published literature. This document also covers advice and best practice in the design and conduct of animal studies to support researchers in improving standards from the start of the experimental design process through to publication.


Asunto(s)
Experimentación Animal , Guías como Asunto , Informe de Investigación , Experimentación Animal/ética , Crianza de Animales Domésticos , Animales , Intervalos de Confianza , Vivienda para Animales , Evaluación de Resultado en la Atención de Salud , Publicaciones , Distribución Aleatoria , Reproducibilidad de los Resultados , Tamaño de la Muestra
14.
BMC Vet Res ; 16(1): 242, 2020 Jul 14.
Artículo en Inglés | MEDLINE | ID: mdl-32660541

RESUMEN

Reproducible science requires transparent reporting. The ARRIVE guidelines (Animal Research: Reporting of In Vivo Experiments) were originally developed in 2010 to improve the reporting of animal research. They consist of a checklist of information to include in publications describing in vivo experiments to enable others to scrutinise the work adequately, evaluate its methodological rigour, and reproduce the methods and results. Despite considerable levels of endorsement by funders and journals over the years, adherence to the guidelines has been inconsistent, and the anticipated improvements in the quality of reporting in animal research publications have not been achieved. Here, we introduce ARRIVE 2.0. The guidelines have been updated and information reorganised to facilitate their use in practice. We used a Delphi exercise to prioritise and divide the items of the guidelines into 2 sets, the "ARRIVE Essential 10," which constitutes the minimum requirement, and the "Recommended Set," which describes the research context. This division facilitates improved reporting of animal research by supporting a stepwise approach to implementation. This helps journal editors and reviewers verify that the most important items are being reported in manuscripts. We have also developed the accompanying Explanation and Elaboration document, which serves (1) to explain the rationale behind each item in the guidelines, (2) to clarify key concepts, and (3) to provide illustrative examples. We aim, through these changes, to help ensure that researchers, reviewers, and journal editors are better equipped to improve the rigour and transparency of the scientific process and thus reproducibility.


Asunto(s)
Experimentación Animal , Guías como Asunto , Informe de Investigación , Animales , Lista de Verificación
15.
Br J Pharmacol ; 177(16): 3617-3624, 2020 08.
Artículo en Inglés | MEDLINE | ID: mdl-32662519

RESUMEN

Reproducible science requires transparent reporting. The ARRIVE guidelines (Animal Research: Reporting of In Vivo Experiments) were originally developed in 2010 to improve the reporting of animal research. They consist of a checklist of information to include in publications describing in vivo experiments to enable others to scrutinise the work adequately, evaluate its methodological rigour, and reproduce the methods and results. Despite considerable levels of endorsement by funders and journals over the years, adherence to the guidelines has been inconsistent, and the anticipated improvements in the quality of reporting in animal research publications have not been achieved. Here, we introduce ARRIVE 2.0. The guidelines have been updated and information reorganised to facilitate their use in practice. We used a Delphi exercise to prioritise and divide the items of the guidelines into 2 sets, the "ARRIVE Essential 10," which constitutes the minimum requirement, and the "Recommended Set," which describes the research context. This division facilitates improved reporting of animal research by supporting a stepwise approach to implementation. This helps journal editors and reviewers verify that the most important items are being reported in manuscripts. We have also developed the accompanying Explanation and Elaboration (E&E) document, which serves (1) to explain the rationale behind each item in the guidelines, (2) to clarify key concepts, and (3) to provide illustrative examples. We aim, through these changes, to help ensure that researchers, reviewers, and journal editors are better equipped to improve the rigour and transparency of the scientific process and thus reproducibility.


Asunto(s)
Experimentación Animal , Animales , Lista de Verificación , Reproducibilidad de los Resultados , Proyectos de Investigación , Informe de Investigación
16.
Exp Physiol ; 105(9): 1459-1466, 2020 09.
Artículo en Inglés | MEDLINE | ID: mdl-32666546

RESUMEN

Reproducible science requires transparent reporting. The ARRIVE guidelines (Animal Research: Reporting of In Vivo Experiments) were originally developed in 2010 to improve the reporting of animal research. They consist of a checklist of information to include in publications describing in vivo experiments to enable others to scrutinise the work adequately, evaluate its methodological rigour, and reproduce the methods and results. Despite considerable levels of endorsement by funders and journals over the years, adherence to the guidelines has been inconsistent, and the anticipated improvements in the quality of reporting in animal research publications have not been achieved. Here, we introduce ARRIVE 2.0. The guidelines have been updated and information reorganised to facilitate their use in practice. We used a Delphi exercise to prioritise and divide the items of the guidelines into 2 sets, the "ARRIVE Essential 10," which constitutes the minimum requirement, and the "Recommended Set," which describes the research context. This division facilitates improved reporting of animal research by supporting a stepwise approach to implementation. This helps journal editors and reviewers verify that the most important items are being reported in manuscripts. We have also developed the accompanying Explanation and Elaboration document, which serves (1) to explain the rationale behind each item in the guidelines, (2) to clarify key concepts, and (3) to provide illustrative examples. We aim, through these changes, to help ensure that researchers, reviewers, and journal editors are better equipped to improve the rigour and transparency of the scientific process and thus reproducibility.


Asunto(s)
Experimentación Animal/normas , Guías como Asunto , Animales , Lista de Verificación , Reproducibilidad de los Resultados , Proyectos de Investigación
17.
J Physiol ; 598(18): 3793-3801, 2020 09.
Artículo en Inglés | MEDLINE | ID: mdl-32666574

RESUMEN

Reproducible science requires transparent reporting. The ARRIVE guidelines (Animal Research: Reporting of In Vivo Experiments) were originally developed in 2010 to improve the reporting of animal research. They consist of a checklist of information to include in publications describing in vivo experiments to enable others to scrutinise the work adequately, evaluate its methodological rigour, and reproduce the methods and results. Despite considerable levels of endorsement by funders and journals over the years, adherence to the guidelines has been inconsistent, and the anticipated improvements in the quality of reporting in animal research publications have not been achieved. Here, we introduce ARRIVE 2.0. The guidelines have been updated and information reorganised to facilitate their use in practice. We used a Delphi exercise to prioritise and divide the items of the guidelines into 2 sets, the 'ARRIVE Essential 10,' which constitutes the minimum requirement, and the 'Recommended Set,' which describes the research context. This division facilitates improved reporting of animal research by supporting a stepwise approach to implementation. This helps journal editors and reviewers verify that the most important items are being reported in manuscripts. We have also developed the accompanying Explanation and Elaboration document, which serves (1) to explain the rationale behind each item in the guidelines, (2) to clarify key concepts, and (3) to provide illustrative examples. We aim, through these changes, to help ensure that researchers, reviewers, and journal editors are better equipped to improve the rigour and transparency of the scientific process and thus reproducibility.


Asunto(s)
Experimentación Animal , Animales , Lista de Verificación , Reproducibilidad de los Resultados , Informe de Investigación
18.
J Cereb Blood Flow Metab ; 40(9): 1769-1777, 2020 09.
Artículo en Inglés | MEDLINE | ID: mdl-32663096

RESUMEN

Reproducible science requires transparent reporting. The ARRIVE guidelines (Animal Research: Reporting of In Vivo Experiments) were originally developed in 2010 to improve the reporting of animal research. They consist of a checklist of information to include in publications describing in vivo experiments to enable others to scrutinise the work adequately, evaluate its methodological rigour, and reproduce the methods and results. Despite considerable levels of endorsement by funders and journals over the years, adherence to the guidelines has been inconsistent, and the anticipated improvements in the quality of reporting in animal research publications have not been achieved. Here, we introduce ARRIVE 2.0. The guidelines have been updated and information reorganised to facilitate their use in practice. We used a Delphi exercise to prioritise and divide the items of the guidelines into 2 sets, the "ARRIVE Essential 10," which constitutes the minimum requirement, and the "Recommended Set," which describes the research context. This division facilitates improved reporting of animal research by supporting a stepwise approach to implementation. This helps journal editors and reviewers verify that the most important items are being reported in manuscripts. We have also developed the accompanying Explanation and Elaboration document, which serves (1) to explain the rationale behind each item in the guidelines, (2) to clarify key concepts, and (3) to provide illustrative examples. We aim, through these changes, to help ensure that researchers, reviewers, and journal editors are better equipped to improve the rigour and transparency of the scientific process and thus reproducibility.

19.
J Cereb Blood Flow Metab ; 40(11): 2152-2164, 2020 11.
Artículo en Inglés | MEDLINE | ID: mdl-32576074

RESUMEN

An important factor for successful translational stroke research is study quality. Low-quality studies are at risk of biased results and effect overestimation, as has been intensely discussed for small animal stroke research. However, little is known about the methodological rigor and quality in large animal stroke models, which are becoming more frequently used in the field. Based on research in two databases, this systematic review surveys and analyses the methodological quality in large animal stroke research. Quality analysis was based on the Stroke Therapy Academic Industry Roundtable and the Animals in Research: Reporting In Vivo Experiments guidelines. Our analysis revealed that large animal models are utilized with similar shortcomings as small animal models. Moreover, translational benefits of large animal models may be limited due to lacking implementation of important quality criteria such as randomization, allocation concealment, and blinded assessment of outcome. On the other hand, an increase of study quality over time and a positive correlation between study quality and journal impact factor were identified. Based on the obtained findings, we derive recommendations for optimal study planning, conducting, and data analysis/reporting when using large animal stroke models to fully benefit from the translational advantages offered by these models.


Asunto(s)
Modelos Animales de Enfermedad , Accidente Cerebrovascular/etiología , Investigación Biomédica Traslacional/métodos , Investigación Biomédica Traslacional/normas , Experimentación Animal/normas , Experimentación Animal/estadística & datos numéricos , Animales , Humanos , Reproducibilidad de los Resultados , Investigación Biomédica Traslacional/estadística & datos numéricos
20.
J Cereb Blood Flow Metab ; 40(7): 1402-1414, 2020 07.
Artículo en Inglés | MEDLINE | ID: mdl-32151228

RESUMEN

Assessment of outcome in preclinical studies of vascular cognitive impairment (VCI) is heterogenous. Through an ARUK Scottish Network supported questionnaire and workshop (mostly UK-based researchers), we aimed to determine underlying variability and what could be implemented to overcome identified challenges. Twelve UK VCI research centres were identified and invited to complete a questionnaire and attend a one-day workshop. Questionnaire responses demonstrated agreement that outcome assessments in VCI preclinical research vary by group and even those common across groups, may be performed differently. From the workshop, six themes were discussed: issues with preclinical models, reasons for choosing functional assessments, issues in interpretation of functional assessments, describing and reporting functional outcome assessments, sharing resources and expertise, and standardization of outcomes. Eight consensus points emerged demonstrating broadly that the chosen assessment should reflect the deficit being measured, and therefore that one assessment does not suit all models; guidance/standardisation on recording VCI outcome reporting is needed and that uniformity would be aided by a platform to share expertise, material, protocols and procedures thus reducing heterogeneity and so increasing potential for collaboration, comparison and replication. As a result of the workshop, UK wide consensus statements were agreed and future priorities for preclinical research identified.


Asunto(s)
Demencia Vascular , Modelos Animales de Enfermedad , Proyectos de Investigación/normas , Animales , Consenso , Recuperación de la Función , Encuestas y Cuestionarios , Reino Unido
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA
...