Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 44
Filtrar
Más filtros

Banco de datos
País/Región como asunto
Tipo del documento
Intervalo de año de publicación
1.
PLoS Biol ; 21(10): e3002362, 2023 Oct.
Artículo en Inglés | MEDLINE | ID: mdl-37856538

RESUMEN

Conversations about open science have reached the mainstream, yet many open science practices such as data sharing remain uncommon. Our efforts towards openness therefore need to increase in scale and aim for a more ambitious target. We need an ecosystem not only where research outputs are openly shared but also in which transparency permeates the research process from the start and lends itself to more rigorous and collaborative research. To support this vision, this Essay provides an overview of a selection of open science initiatives from the past 2 decades, focusing on methods transparency, scholarly communication, team science, and research culture, and speculates about what the future of open science could look like. It then draws on these examples to provide recommendations for how funders, institutions, journals, regulators, and other stakeholders can create an environment that is ripe for improvement.


Asunto(s)
Comunicación , Ecosistema , Difusión de la Información , Comunicación Académica
2.
PLoS Biol ; 21(6): e3002133, 2023 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-37390046

RESUMEN

Characterizing cellular diversity at different levels of biological organization and across data modalities is a prerequisite to understanding the function of cell types in the brain. Classification of neurons is also essential to manipulate cell types in controlled ways and to understand their variation and vulnerability in brain disorders. The BRAIN Initiative Cell Census Network (BICCN) is an integrated network of data-generating centers, data archives, and data standards developers, with the goal of systematic multimodal brain cell type profiling and characterization. Emphasis of the BICCN is on the whole mouse brain with demonstration of prototype feasibility for human and nonhuman primate (NHP) brains. Here, we provide a guide to the cellular and spatial approaches employed by the BICCN, and to accessing and using these data and extensive resources, including the BRAIN Cell Data Center (BCDC), which serves to manage and integrate data across the ecosystem. We illustrate the power of the BICCN data ecosystem through vignettes highlighting several BICCN analysis and visualization tools. Finally, we present emerging standards that have been developed or adopted toward Findable, Accessible, Interoperable, and Reusable (FAIR) neuroscience. The combined BICCN ecosystem provides a comprehensive resource for the exploration and analysis of cell types in the brain.


Asunto(s)
Encéfalo , Neurociencias , Animales , Humanos , Ratones , Ecosistema , Neuronas
3.
Nucleic Acids Res ; 51(D1): D358-D367, 2023 01 06.
Artículo en Inglés | MEDLINE | ID: mdl-36370112

RESUMEN

Antibodies are ubiquitous key biological research resources yet are tricky to use as they are prone to performance issues and represent a major source of variability across studies. Understanding what antibody was used in a published study is therefore necessary to repeat and/or interpret a given study. However, antibody reagents are still frequently not cited with sufficient detail to determine which antibody was used in experiments. The Antibody Registry is a public, open database that enables citation of antibodies by providing a persistent record for any antibody-based reagent used in a publication. The registry is the authority for antibody Research Resource Identifiers, or RRIDs, which are requested or required by hundreds of journals seeking to improve the citation of these key resources. The registry is the most comprehensive listing of persistently identified antibody reagents used in the scientific literature. Data contributors span individual authors who use antibodies to antibody companies, which provide their entire catalogs including discontinued items. Unlike many commercial antibody listing sites which tend to remove reagents no longer sold, registry records persist, providing an interface between a fast-moving commercial marketplace and the static scientific literature. The Antibody Registry (RRID:SCR_006397) https://antibodyregistry.org.


Asunto(s)
Anticuerpos , Bases de Datos Factuales , Sistema de Registros
4.
PLoS Comput Biol ; 17(5): e1008967, 2021 05.
Artículo en Inglés | MEDLINE | ID: mdl-34043624

RESUMEN

Antibodies are widely used reagents to test for expression of proteins and other antigens. However, they might not always reliably produce results when they do not specifically bind to the target proteins that their providers designed them for, leading to unreliable research results. While many proposals have been developed to deal with the problem of antibody specificity, it is still challenging to cover the millions of antibodies that are available to researchers. In this study, we investigate the feasibility of automatically generating alerts to users of problematic antibodies by extracting statements about antibody specificity reported in the literature. The extracted alerts can be used to construct an "Antibody Watch" knowledge base containing supporting statements of problematic antibodies. We developed a deep neural network system and tested its performance with a corpus of more than two thousand articles that reported uses of antibodies. We divided the problem into two tasks. Given an input article, the first task is to identify snippets about antibody specificity and classify if the snippets report that any antibody exhibits non-specificity, and thus is problematic. The second task is to link each of these snippets to one or more antibodies mentioned in the snippet. The experimental evaluation shows that our system can accurately perform the classification task with 0.925 weighted F1-score, linking with 0.962 accuracy, and 0.914 weighted F1 when combined to complete the joint task. We leveraged Research Resource Identifiers (RRID) to precisely identify antibodies linked to the extracted specificity snippets. The result shows that it is feasible to construct a reliable knowledge base about problematic antibodies by text mining.


Asunto(s)
Especificidad de Anticuerpos , Minería de Datos , Animales , Humanos , Ratones , Redes Neurales de la Computación
5.
J Med Internet Res ; 24(6): e37324, 2022 06 27.
Artículo en Inglés | MEDLINE | ID: mdl-35759334

RESUMEN

BACKGROUND: Improving rigor and transparency measures should lead to improvements in reproducibility across the scientific literature; however, the assessment of measures of transparency tends to be very difficult if performed manually. OBJECTIVE: This study addresses the enhancement of the Rigor and Transparency Index (RTI, version 2.0), which attempts to automatically assess the rigor and transparency of journals, institutions, and countries using manuscripts scored on criteria found in reproducibility guidelines (eg, Materials Design, Analysis, and Reporting checklist criteria). METHODS: The RTI tracks 27 entity types using natural language processing techniques such as Bidirectional Long Short-term Memory Conditional Random Field-based models and regular expressions; this allowed us to assess over 2 million papers accessed through PubMed Central. RESULTS: Between 1997 and 2020 (where data were readily available in our data set), rigor and transparency measures showed general improvement (RTI 2.29 to 4.13), suggesting that authors are taking the need for improved reporting seriously. The top-scoring journals in 2020 were the Journal of Neurochemistry (6.23), British Journal of Pharmacology (6.07), and Nature Neuroscience (5.93). We extracted the institution and country of origin from the author affiliations to expand our analysis beyond journals. Among institutions publishing >1000 papers in 2020 (in the PubMed Central open access set), Capital Medical University (4.75), Yonsei University (4.58), and University of Copenhagen (4.53) were the top performers in terms of RTI. In country-level performance, we found that Ethiopia and Norway consistently topped the RTI charts of countries with 100 or more papers per year. In addition, we tested our assumption that the RTI may serve as a reliable proxy for scientific replicability (ie, a high RTI represents papers containing sufficient information for replication efforts). Using work by the Reproducibility Project: Cancer Biology, we determined that replication papers (RTI 7.61, SD 0.78) scored significantly higher (P<.001) than the original papers (RTI 3.39, SD 1.12), which according to the project required additional information from authors to begin replication efforts. CONCLUSIONS: These results align with our view that RTI may serve as a reliable proxy for scientific replicability. Unfortunately, RTI measures for journals, institutions, and countries fall short of the replicated paper average. If we consider the RTI of these replication studies as a target for future manuscripts, more work will be needed to ensure that the average manuscript contains sufficient information for replication attempts.


Asunto(s)
Lista de Verificación , Edición , Humanos , Noruega , Reproducibilidad de los Resultados , Proyectos de Investigación
6.
J Neurosci Res ; 97(4): 377-390, 2019 04.
Artículo en Inglés | MEDLINE | ID: mdl-30506706

RESUMEN

Progress in basic and clinical research is slowed when researchers fail to provide a complete and accurate report of how a study was designed, executed, and the results analyzed. Publishing rigorous scientific research involves a full description of the methods, materials, procedures, and outcomes. Investigators may fail to provide a complete description of how their study was designed and executed because they may not know how to accurately report the information or the mechanisms are not in place to facilitate transparent reporting. Here, we provide an overview of how authors can write manuscripts in a transparent and thorough manner. We introduce a set of reporting criteria that can be used for publishing, including recommendations on reporting the experimental design and statistical approaches. We also discuss how to accurately visualize the results and provide recommendations for peer reviewers to enhance rigor and transparency. Incorporating transparency practices into research manuscripts will significantly improve the reproducibility of the results by independent laboratories.


Asunto(s)
Investigación Biomédica/normas , Edición/normas , Exactitud de los Datos , Humanos , Mejoramiento de la Calidad , Reproducibilidad de los Resultados , Proyectos de Investigación/normas
7.
Immunology ; 147(3): 349-54, 2016 Mar.
Artículo en Inglés | MEDLINE | ID: mdl-26678806

RESUMEN

The Immune Epitope Database is uniquely positioned to assess the body of research related to immune epitopes, we have manually curated all such published data. Thus, we are able to make observations on the state of these fields of research, as well as aggregate the individual data points to present a clearer picture of the immune response to specific antigens in all studied hosts. Additionally, we are able to identify where conflicts in the literature exist and where publications fall short in terms of identifiable methods and in reproducibility. Here we present guidelines to improve the quality of immune epitope data, which will benefit journals and researchers alike.


Asunto(s)
Alergia e Inmunología/normas , Bases de Datos de Proteínas/normas , Epítopos/inmunología , Humanos , Reproducibilidad de los Resultados
9.
bioRxiv ; 2024 Jan 16.
Artículo en Inglés | MEDLINE | ID: mdl-38293091

RESUMEN

Research resources like transgenic animals and antibodies are the workhorses of biomedicine, enabling investigators to relatively easily study specific disease conditions. As key biological resources, transgenic animals and antibodies are often validated, maintained, and distributed from university based stock centers. As these centers heavily rely largely on grant funding, it is critical that they are cited by investigators so that usage can be tracked. However, unlike systems for tracking the impact of papers, the conventions and systems for tracking key resource usage and impact lag behind. Previous studies have shown that about 50% of the resources are not findable, making the studies they are supporting irreproducible, but also makes tracking resources difficult. The RRID project is filling this gap by working with journals and resource providers to improve citation practices and to track the usage of these key resources. Here, we reviewed 10 years of citation practices for five university based stock centers, characterizing each reference into two broad categories: findable (authors could use the RRID, stock number, or full name) and not findable (authors could use a nickname or a common name that is not unique to the resource). The data revealed that when stock centers asked their communities to cite resources by RRID, in addition to helping stock centers more easily track resource usage by increasing the number of RRID papers, authors shifted from citing resources predominantly by nickname (~50% of the time) to citing them by one of the findable categories (~85%) in a matter of several years. In the case of one stock center, the MMRRC, the improvement in findability is also associated with improvements in the adherence to NIH rigor criteria, as determined by a significant increase in the Rigor and Transparency Index for studies using MMRRC mice. From this data, it was not possible to determine whether outreach to authors or changes to stock center websites drove better citation practices, but findability of research resources and rigor adherence was improved.

10.
Animals (Basel) ; 14(11)2024 May 23.
Artículo en Inglés | MEDLINE | ID: mdl-38891588

RESUMEN

The documentation, preservation and rescue of biological diversity increasingly uses living biological samples. Persistent associations between species, biosamples, such as tissues and cell lines, and the accompanying data are indispensable for using, exchanging and benefiting from these valuable materials. Explicit authentication of such biosamples by assigning unique and robust identifiers is therefore required to allow for unambiguous referencing, avoid identification conflicts and maintain reproducibility in research. A predefined nomenclature based on uniform rules would facilitate this process. However, such a nomenclature is currently lacking for animal biological material. We here present a first, standardized, human-readable nomenclature design, which is sufficient to generate unique and stable identifying names for animal cellular material with a focus on wildlife species. A species-specific human- and machine-readable syntax is included in the proposed standard naming scheme, allowing for the traceability of donated material and cultured cells, as well as data FAIRification. Only when it is consistently applied in the public domain, as publications and inter-institutional samples and data are exchanged, distributed and stored centrally, can the risks of misidentification and loss of traceability be mitigated. This innovative globally applicable identification system provides a standard for a sustainable structure for the long-term storage of animal bio-samples in cryobanks and hence facilitates current as well as future species conservation and biomedical research.

11.
PLoS One ; 18(3): e0281659, 2023.
Artículo en Inglés | MEDLINE | ID: mdl-36888577

RESUMEN

Preprints, versions of scientific manuscripts that precede peer review, are growing in popularity. They offer an opportunity to democratize and accelerate research, as they have no publication costs or a lengthy peer review process. Preprints are often later published in peer-reviewed venues, but these publications and the original preprints are frequently not linked in any way. To this end, we developed a tool, PreprintMatch, to find matches between preprints and their corresponding published papers, if they exist. This tool outperforms existing techniques to match preprints and papers, both on matching performance and speed. PreprintMatch was applied to search for matches between preprints (from bioRxiv and medRxiv), and PubMed. The preliminary nature of preprints offers a unique perspective into scientific projects at a relatively early stage, and with better matching between preprint and paper, we explored questions related to research inequity. We found that preprints from low income countries are published as peer-reviewed papers at a lower rate than high income countries (39.6% and 61.1%, respectively), and our data is consistent with previous work that cite a lack of resources, lack of stability, and policy choices to explain this discrepancy. Preprints from low income countries were also found to be published quicker (178 vs 203 days) and with less title, abstract, and author similarity to the published version compared to high income countries. Low income countries add more authors from the preprint to the published version than high income countries (0.42 authors vs 0.32, respectively), a practice that is significantly more frequent in China compared to similar countries. Finally, we find that some publishers publish work with authors from lower income countries more frequently than others.


Asunto(s)
Revisión por Pares , PubMed , China
12.
bioRxiv ; 2023 Aug 17.
Artículo en Inglés | MEDLINE | ID: mdl-37398479

RESUMEN

Antibodies are critical reagents to detect and characterize proteins. It is commonly understood that many commercial antibodies do not recognize their intended targets, but information on the scope of the problem remains largely anecdotal, and as such, feasibility of the goal of at least one potent and specific antibody targeting each protein in a proteome cannot be assessed. Focusing on antibodies for human proteins, we have scaled a standardized characterization approach using parental and knockout cell lines (Laflamme et al., 2019) to assess the performance of 614 commercial antibodies for 65 neuroscience-related proteins. Side-by-side comparisons of all antibodies against each target, obtained from multiple commercial partners, demonstrates that: i) more than 50% of all antibodies failed in one or more tests, ii) yet, ~50-75% of the protein set was covered by at least one high-performing antibody, depending on application, suggesting that coverage of human proteins by commercial antibodies is significant; and iii) recombinant antibodies performed better than monoclonal or polyclonal antibodies. The hundreds of underperforming antibodies identified in this study were found to have been used in a large number of published articles, which should raise alarm. Encouragingly, more than half of the underperforming commercial antibodies were reassessed by the manufacturers, and many had alterations to their recommended usage or were removed from the market. This first such study helps demonstrate the scale of the antibody specificity problem but also suggests an efficient strategy toward achieving coverage of the human proteome; mine the existing commercial antibody repertoire, and use the data to focus new renewable antibody generation efforts.

13.
Elife ; 122023 Nov 23.
Artículo en Inglés | MEDLINE | ID: mdl-37995198

RESUMEN

Antibodies are critical reagents to detect and characterize proteins. It is commonly understood that many commercial antibodies do not recognize their intended targets, but information on the scope of the problem remains largely anecdotal, and as such, feasibility of the goal of at least one potent and specific antibody targeting each protein in a proteome cannot be assessed. Focusing on antibodies for human proteins, we have scaled a standardized characterization approach using parental and knockout cell lines (Laflamme et al., 2019) to assess the performance of 614 commercial antibodies for 65 neuroscience-related proteins. Side-by-side comparisons of all antibodies against each target, obtained from multiple commercial partners, have demonstrated that: (i) more than 50% of all antibodies failed in one or more applications, (ii) yet, ~50-75% of the protein set was covered by at least one high-performing antibody, depending on application, suggesting that coverage of human proteins by commercial antibodies is significant; and (iii) recombinant antibodies performed better than monoclonal or polyclonal antibodies. The hundreds of underperforming antibodies identified in this study were found to have been used in a large number of published articles, which should raise alarm. Encouragingly, more than half of the underperforming commercial antibodies were reassessed by the manufacturers, and many had alterations to their recommended usage or were removed from the market. This first study helps demonstrate the scale of the antibody specificity problem but also suggests an efficient strategy toward achieving coverage of the human proteome; mine the existing commercial antibody repertoire, and use the data to focus new renewable antibody generation efforts.


Commercially produced antibodies are essential research tools. Investigators at universities and pharmaceutical companies use them to study human proteins, which carry out all the functions of the cells. Scientists usually buy antibodies from commercial manufacturers who produce more than 6 million antibody products altogether. Yet many commercial antibodies do not work as advertised. They do not recognize their intended protein target or may flag untargeted proteins. Both can skew research results and make it challenging to reproduce scientific studies, which is vital to scientific integrity. Using ineffective commercial antibodies likely wastes $1 billion in research funding each year. Large-scale validation of commercial antibodies by an independent third party could reduce the waste and misinformation associated with using ineffective commercial antibodies. Previous research testing an antibody validation pipeline showed that a commercial antibody widely used in studies to detect a protein involved in amyotrophic lateral sclerosis did not work. Meanwhile, the best-performing commercial antibodies were not used in research. Testing commercial antibodies and making the resulting data available would help scientists identify the best study tools and improve research reliability. Ayoubi et al. collaborated with antibody manufacturers and organizations that produce genetic knock-out cell lines to develop a system validating the effectiveness of commercial antibodies. In the experiments, Ayoubi et al. tested 614 commercial antibodies intended to detect 65 proteins involved in neurologic diseases. An effective antibody was available for about two thirds of the 65 proteins. Yet, hundreds of the antibodies, including many used widely in studies, were ineffective. Manufacturers removed some underperforming antibodies from the market or altered their recommended uses based on these data. Ayoubi et al. shared the resulting data on Zenodo, a publicly available preprint database. The experiments suggest that 20-30% of protein studies use ineffective antibodies, indicating a substantial need for independent assessment of commercial antibodies. Ayoubi et al. demonstrated their side-by-side antibody comparison methods were an effective and efficient way of validating commercial antibodies. Using this approach to test commercial antibodies against all human proteins would cost about $50 million. But it could save much of the $1 billion wasted each year on research involving ineffective antibodies. Independent validation of commercial antibodies could also reduce wasted efforts by scientists using ineffective antibodies and improve the reliability of research results. It would also enable faster, more reliable research that may help scientists understand diseases and develop new therapies to improve patient's lives.


Asunto(s)
Anticuerpos , Proteoma , Humanos , Anticuerpos/química
14.
Gigascience ; 112022 06 14.
Artículo en Inglés | MEDLINE | ID: mdl-35701373

RESUMEN

Research resource identifiers (RRIDs) are persistent unique identifiers for scientific resources used to conduct studies such as reagents and tools. Inclusion of these identifiers into the scientific literature has been demonstrated to improve the reproducibility of papers because resources, like antibodies, are easier to find, making methods easier to reproduce. RRIDs also dramatically reduce the use of problematic resources, such as contaminated cell lines. The addition of RRIDs into a manuscript means that authors have to look up information that they may have previously omitted or confront information about problems that may have been reported about their resources. The use of RRIDs is primarily driven by champion journals, such as GigaScience and others. Although still nascent, this practice lays important groundwork for citation types that can cover non-traditional scholarly output, such as software tools and key reagents; giving authors of various types of tools scholarly credit for their contributions.


Asunto(s)
Aprendizaje , Programas Informáticos , Reproducibilidad de los Resultados
15.
BMC Res Notes ; 15(1): 203, 2022 Jun 11.
Artículo en Inglés | MEDLINE | ID: mdl-35690782

RESUMEN

The rising rate of preprints and publications, combined with persistent inadequate reporting practices and problems with study design and execution, have strained the traditional peer review system. Automated screening tools could potentially enhance peer review by helping authors, journal editors, and reviewers to identify beneficial practices and common problems in preprints or submitted manuscripts. Tools can screen many papers quickly, and may be particularly helpful in assessing compliance with journal policies and with straightforward items in reporting guidelines. However, existing tools cannot understand or interpret the paper in the context of the scientific literature. Tools cannot yet determine whether the methods used are suitable to answer the research question, or whether the data support the authors' conclusions. Editors and peer reviewers are essential for assessing journal fit and the overall quality of a paper, including the experimental design, the soundness of the study's conclusions, potential impact and innovation. Automated screening tools cannot replace peer review, but may aid authors, reviewers, and editors in improving scientific papers. Strategies for responsible use of automated tools in peer review may include setting performance criteria for tools, transparently reporting tool performance and use, and training users to interpret reports.


Asunto(s)
Políticas Editoriales , Revisión de la Investigación por Pares , Proyectos de Investigación , Informe de Investigación
16.
N Biotechnol ; 65: 1-8, 2021 Nov 25.
Artículo en Inglés | MEDLINE | ID: mdl-34246180

RESUMEN

A vast array of commercial antibodies covers a large percentage of human gene products, but determining which among them is most appropriate for any given application is challenging. This leads to use of non-specific antibodies that contributes to issues with reproducibility. It is our opinion that the community of scientists who use commercial antibodies in their biomedical research would benefit from third-party antibody characterization entities that use standardized operating procedures to assess and compare antibody performance. Ideally, such entities would follow the principles of open science, such that all antibodies against any given protein target would be tested in parallel, and all data generated released to the public domain without bias. Furthermore, there should be no financial incentive for the entity beyond cost-recovery. Such non-profit organizations, combined with other scientific efforts, could catalyse new discoveries by providing scientists with better validated antibody tools.


Asunto(s)
Anticuerpos , Investigación Biomédica , Indicadores y Reactivos/normas , Humanos , Reproducibilidad de los Resultados
17.
Front Physiol ; 12: 693735, 2021.
Artículo en Inglés | MEDLINE | ID: mdl-34248680

RESUMEN

The Data and Resource Center (DRC) of the NIH-funded SPARC program is developing databases, connectivity maps, and simulation tools for the mammalian autonomic nervous system. The experimental data and mathematical models supplied to the DRC by the SPARC consortium are curated, annotated and semantically linked via a single knowledgebase. A data portal has been developed that allows discovery of data and models both via semantic search and via an interface that includes Google Map-like 2D flatmaps for displaying connectivity, and 3D anatomical organ scaffolds that provide a common coordinate framework for cross-species comparisons. We discuss examples that illustrate the data pipeline, which includes data upload, curation, segmentation (for image data), registration against the flatmaps and scaffolds, and finally display via the web portal, including the link to freely available online computational facilities that will enable neuromodulation hypotheses to be investigated by the autonomic neuroscience community and device manufacturers.

18.
Genes Brain Behav ; : e12738, 2021 Apr 23.
Artículo en Inglés | MEDLINE | ID: mdl-33893716

RESUMEN

The National Institute on Drug Abuse and Joint Institute for Biological Sciences at the Oak Ridge National Laboratory hosted a meeting attended by a diverse group of scientists with expertise in substance use disorders (SUDs), computational biology, and FAIR (Findability, Accessibility, Interoperability, and Reusability) data sharing. The meeting's objective was to discuss and evaluate better strategies to integrate genetic, epigenetic, and 'omics data across human and model organisms to achieve deeper mechanistic insight into SUDs. Specific topics were to (a) evaluate the current state of substance use genetics and genomics research and fundamental gaps, (b) identify opportunities and challenges of integration and sharing across species and data types, (c) identify current tools and resources for integration of genetic, epigenetic, and phenotypic data, (d) discuss steps and impediment related to data integration, and (e) outline future steps to support more effective collaboration-particularly between animal model research communities and human genetics and clinical research teams. This review summarizes key facets of this catalytic discussion with a focus on new opportunities and gaps in resources and knowledge on SUDs.

19.
Database (Oxford) ; 20202020 01 01.
Artículo en Inglés | MEDLINE | ID: mdl-31925435

RESUMEN

The ever accelerating pace of biomedical research results in corresponding acceleration in the volume of biomedical literature created. Since new research builds upon existing knowledge, the rate of increase in the available knowledge encoded in biomedical literature makes the easy access to that implicit knowledge more vital over time. Toward the goal of making implicit knowledge in the biomedical literature easily accessible to biomedical researchers, we introduce a question answering system called Bio-AnswerFinder. Bio-AnswerFinder uses a weighted-relaxed word mover's distance based similarity on word/phrase embeddings learned from PubMed abstracts to rank answers after question focus entity type filtering. Our approach retrieves relevant documents iteratively via enhanced keyword queries from a traditional search engine. To improve document retrieval performance, we introduced a supervised long short term memory neural network to select keywords from the question to facilitate iterative keyword search. Our unsupervised baseline system achieves a mean reciprocal rank score of 0.46 and Precision@1 of 0.32 on 936 questions from BioASQ. The answer sentences are further ranked by a fine-tuned bidirectional encoder representation from transformers (BERT) classifier trained using 100 answer candidate sentences per question for 492 BioASQ questions. To test ranking performance, we report a blind test on 100 questions that three independent annotators scored. These experts preferred BERT based reranking with 7% improvement on MRR and 13% improvement on Precision@1 scores on average.


Asunto(s)
Investigación Biomédica , Almacenamiento y Recuperación de la Información/métodos , Procesamiento de Lenguaje Natural , Redes Neurales de la Computación , Minería de Datos , Humanos
20.
iScience ; 23(11): 101698, 2020 Nov 20.
Artículo en Inglés | MEDLINE | ID: mdl-33196023

RESUMEN

The reproducibility crisis is a multifaceted problem involving ingrained practices within the scientific community. Fortunately, some causes are addressed by the author's adherence to rigor and reproducibility criteria, implemented via checklists at various journals. We developed an automated tool (SciScore) that evaluates research articles based on their adherence to key rigor criteria, including NIH criteria and RRIDs, at an unprecedented scale. We show that despite steady improvements, less than half of the scoring criteria, such as blinding or power analysis, are routinely addressed by authors; digging deeper, we examined the influence of specific checklists on average scores. The average score for a journal in a given year was named the Rigor and Transparency Index (RTI), a new journal quality metric. We compared the RTI with the Journal Impact Factor and found there was no correlation. The RTI can potentially serve as a proxy for methodological quality.

SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA