Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 172
Filtrar
Mais filtros

Intervalo de ano de publicação
1.
Physiol Rev ; 104(3): 1387-1408, 2024 Jul 01.
Artigo em Inglês | MEDLINE | ID: mdl-38451234

RESUMO

Effective data management is crucial for scientific integrity and reproducibility, a cornerstone of scientific progress. Well-organized and well-documented data enable validation and building on results. Data management encompasses activities including organization, documentation, storage, sharing, and preservation. Robust data management establishes credibility, fostering trust within the scientific community and benefiting researchers' careers. In experimental biomedicine, comprehensive data management is vital due to the typically intricate protocols, extensive metadata, and large datasets. Low-throughput experiments, in particular, require careful management to address variations and errors in protocols and raw data quality. Transparent and accountable research practices rely on accurate documentation of procedures, data collection, and analysis methods. Proper data management ensures long-term preservation and accessibility of valuable datasets. Well-managed data can be revisited, contributing to cumulative knowledge and potential new discoveries. Publicly funded research has an added responsibility for transparency, resource allocation, and avoiding redundancy. Meeting funding agency expectations increasingly requires rigorous methodologies, adherence to standards, comprehensive documentation, and widespread sharing of data, code, and other auxiliary resources. This review provides critical insights into raw and processed data, metadata, high-throughput versus low-throughput datasets, a common language for documentation, experimental and reporting guidelines, efficient data management systems, sharing practices, and relevant repositories. We systematically present available resources and optimal practices for wide use by experimental biomedical researchers.


Assuntos
Pesquisa Biomédica , Gerenciamento de Dados , Disseminação de Informação , Pesquisa Biomédica/normas , Pesquisa Biomédica/métodos , Disseminação de Informação/métodos , Humanos , Animais , Gerenciamento de Dados/métodos
2.
BMC Bioinformatics ; 25(1): 210, 2024 Jun 12.
Artigo em Inglês | MEDLINE | ID: mdl-38867185

RESUMO

BACKGROUND: In the realm of biomedical research, the growing volume, diversity and quantity of data has escalated the demand for statistical analysis as it is indispensable for synthesizing, interpreting, and publishing data. Hence the need for accessible analysis tools drastically increased. StatiCAL emerges as a user-friendly solution, enabling researchers to conduct basic analyses without necessitating extensive programming expertise. RESULTS: StatiCAL includes divers functionalities: data management, visualization on variables and statistical analysis. Data management functionalities allow the user to freely add or remove variables, to select sub-population and to visualise selected data to better perform the analysis. With this tool, users can freely perform statistical analysis such as descriptive, graphical, univariate, and multivariate analysis. All of this can be performed without the need to learn R coding as the software is a graphical user interface where all the action can be performed by clicking a button. CONCLUSIONS: StatiCAL represents a valuable contribution to the field of biomedical research. By being open-access and by providing an intuitive interface with robust features, StatiCAL allow researchers to gain autonomy in conducting their projects.


Assuntos
Pesquisa Biomédica , Software , Interface Usuário-Computador , Biologia Computacional/métodos , Gerenciamento de Dados/métodos , Interpretação Estatística de Dados
3.
Nat Methods ; 18(3): 262-271, 2021 03.
Artigo em Inglês | MEDLINE | ID: mdl-33649586

RESUMO

Single-cell technologies have made it possible to profile millions of cells, but for these resources to be useful they must be easy to query and access. To facilitate interactive and intuitive access to single-cell data we have developed scfind, a single-cell analysis tool that facilitates fast search of biologically or clinically relevant marker genes in cell atlases. Using transcriptome data from six mouse cell atlases, we show how scfind can be used to evaluate marker genes, perform in silico gating, and identify both cell-type-specific and housekeeping genes. Moreover, we have developed a subquery optimization routine to ensure that long and complex queries return meaningful results. To make scfind more user friendly, we use indices of PubMed abstracts and techniques from natural language processing to allow for arbitrary queries. Finally, we show how scfind can be used for multi-omics analyses by combining single-cell ATAC-seq data with transcriptome data.


Assuntos
Gerenciamento de Dados/métodos , Armazenamento e Recuperação da Informação/métodos , Análise de Célula Única/métodos , Transcriptoma/genética , Algoritmos , Animais , Análise de Dados , Bases de Dados Genéticas , Regulação da Expressão Gênica , Camundongos , Processamento de Linguagem Natural , PubMed , Interface Usuário-Computador
4.
PLoS Biol ; 19(3): e3001129, 2021 03.
Artigo em Inglês | MEDLINE | ID: mdl-33770077

RESUMO

Decades of reductionist approaches in biology have achieved spectacular progress, but the proliferation of subdisciplines, each with its own technical and social practices regarding data, impedes the growth of the multidisciplinary and interdisciplinary approaches now needed to address pressing societal challenges. Data integration is key to a reintegrated biology able to address global issues such as climate change, biodiversity loss, and sustainable ecosystem management. We identify major challenges to data integration and present a vision for a "Data as a Service"-oriented architecture to promote reuse of data for discovery. The proposed architecture includes standards development, new tools and services, and strategies for career-development and sustainability.


Assuntos
Gerenciamento de Dados/métodos , Disseminação de Informação/métodos , Pesquisa Interdisciplinar/tendências , Biodiversidade , Disciplinas das Ciências Biológicas , Conservação dos Recursos Naturais , Ecossistema , Comunicação Interdisciplinar , Pesquisa Interdisciplinar/métodos
5.
PLoS Biol ; 19(4): e3001162, 2021 04.
Artigo em Inglês | MEDLINE | ID: mdl-33872298

RESUMO

Many randomized controlled trials (RCTs) are biased and difficult to reproduce due to methodological flaws and poor reporting. There is increasing attention for responsible research practices and implementation of reporting guidelines, but whether these efforts have improved the methodological quality of RCTs (e.g., lower risk of bias) is unknown. We, therefore, mapped risk-of-bias trends over time in RCT publications in relation to journal and author characteristics. Meta-information of 176,620 RCTs published between 1966 and 2018 was extracted. The risk-of-bias probability (random sequence generation, allocation concealment, blinding of patients/personnel, and blinding of outcome assessment) was assessed using a risk-of-bias machine learning tool. This tool was simultaneously validated using 63,327 human risk-of-bias assessments obtained from 17,394 RCTs evaluated in the Cochrane Database of Systematic Reviews (CDSR). Moreover, RCT registration and CONSORT Statement reporting were assessed using automated searches. Publication characteristics included the number of authors, journal impact factor (JIF), and medical discipline. The annual number of published RCTs substantially increased over 4 decades, accompanied by increases in authors (5.2 to 7.8) and institutions (2.9 to 4.8). The risk of bias remained present in most RCTs but decreased over time for allocation concealment (63% to 51%), random sequence generation (57% to 36%), and blinding of outcome assessment (58% to 52%). Trial registration (37% to 47%) and the use of the CONSORT Statement (1% to 20%) also rapidly increased. In journals with a higher impact factor (>10), the risk of bias was consistently lower with higher levels of RCT registration and the use of the CONSORT Statement. Automated risk-of-bias predictions had accuracies above 70% for allocation concealment (70.7%), random sequence generation (72.1%), and blinding of patients/personnel (79.8%), but not for blinding of outcome assessment (62.7%). In conclusion, the likelihood of bias in RCTs has generally decreased over the last decades. This optimistic trend may be driven by increased knowledge augmented by mandatory trial registration and more stringent reporting guidelines and journal requirements. Nevertheless, relatively high probabilities of bias remain, particularly in journals with lower impact factors. This emphasizes that further improvement of RCT registration, conduct, and reporting is still urgently needed.


Assuntos
Publicações , Ensaios Clínicos Controlados Aleatórios como Assunto/métodos , Ensaios Clínicos Controlados Aleatórios como Assunto/estatística & dados numéricos , Viés , Bibliometria , Confiabilidade dos Dados , Gerenciamento de Dados/história , Gerenciamento de Dados/métodos , Gerenciamento de Dados/normas , Gerenciamento de Dados/tendências , Bases de Dados Bibliográficas/história , Bases de Dados Bibliográficas/normas , Bases de Dados Bibliográficas/tendências , História do Século XX , História do Século XXI , Humanos , Avaliação de Resultados em Cuidados de Saúde , Registros Públicos de Dados de Cuidados de Saúde , Publicações/história , Publicações/normas , Publicações/estatística & dados numéricos , Publicações/tendências , Melhoria de Qualidade/história , Melhoria de Qualidade/tendências , Ensaios Clínicos Controlados Aleatórios como Assunto/história , Revisões Sistemáticas como Assunto
6.
J Microsc ; 294(3): 350-371, 2024 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-38752662

RESUMO

Bioimage data are generated in diverse research fields throughout the life and biomedical sciences. Its potential for advancing scientific progress via modern, data-driven discovery approaches reaches beyond disciplinary borders. To fully exploit this potential, it is necessary to make bioimaging data, in general, multidimensional microscopy images and image series, FAIR, that is, findable, accessible, interoperable and reusable. These FAIR principles for research data management are now widely accepted in the scientific community and have been adopted by funding agencies, policymakers and publishers. To remain competitive and at the forefront of research, implementing the FAIR principles into daily routines is an essential but challenging task for researchers and research infrastructures. Imaging core facilities, well-established providers of access to imaging equipment and expertise, are in an excellent position to lead this transformation in bioimaging research data management. They are positioned at the intersection of research groups, IT infrastructure providers, the institution´s administration, and microscope vendors. In the frame of German BioImaging - Society for Microscopy and Image Analysis (GerBI-GMB), cross-institutional working groups and third-party funded projects were initiated in recent years to advance the bioimaging community's capability and capacity for FAIR bioimage data management. Here, we provide an imaging-core-facility-centric perspective outlining the experience and current strategies in Germany to facilitate the practical adoption of the FAIR principles closely aligned with the international bioimaging community. We highlight which tools and services are ready to be implemented and what the future directions for FAIR bioimage data have to offer.


Assuntos
Microscopia , Pesquisa Biomédica/métodos , Gerenciamento de Dados/métodos , Processamento de Imagem Assistida por Computador/métodos , Microscopia/métodos
7.
J Med Libr Assoc ; 112(2): 142-144, 2024 Apr 01.
Artigo em Inglês | MEDLINE | ID: mdl-39119154

RESUMO

The DMPTool NIH Data Management and Sharing Plan (DMSP) Templates Project was launched in response to the 2023 NIH Data Management and Sharing (DMS) Policy. This new policy introduced a more structured framework for DMS Plans, featuring six key elements, a departure from the 2003 NIH DMS policy. The project aimed to simplify the process for data librarians, research administrators, and researchers by providing a template with curated guidance, eliminating the need to navigate various policies and guidelines. The template breaks out each Plan section and subsection and provides related guidance and examples at the point of need. This effort has resulted in two NIH DMSP Templates. The first is a generic template (NIH-Default) for all ICs, complying with NOT-OD-21-013 and NOT-OD-22-198. More recently, an NIMH-specific template (NIH-NIMH) was added based on NOT-MH-23-100. As of October 2023, over 5,000 DMS Plans have been written using the main NIH-Default template and the NIH-NIMH alternative template.


Assuntos
National Institutes of Health (U.S.) , Estados Unidos , National Institutes of Health (U.S.)/organização & administração , Humanos , Disseminação de Informação/métodos , Gerenciamento de Dados/métodos
8.
J Med Libr Assoc ; 112(1): 42-47, 2024 Jan 16.
Artigo em Inglês | MEDLINE | ID: mdl-38911529

RESUMO

Background: By defining search strategies and related database exports as code/scripts and data, librarians and information professionals can expand the mandate of research data management (RDM) infrastructure to include this work. This new initiative aimed to create a space in McGill University's institutional data repository for our librarians to deposit and share their search strategies for knowledge syntheses (KS). Case Presentation: The authors, a health sciences librarian and an RDM specialist, created a repository collection of librarian-authored knowledge synthesis (KS) searches in McGill University's Borealis Dataverse collection. We developed and hosted a half-day "Dataverse-a-thon" where we worked with a team of health sciences librarians to develop a standardized KS data management plan (DMP), search reporting documentation, Dataverse software training, and howto guidance for the repository. Conclusion: In addition to better documentation and tracking of KS searches at our institution, the KS Dataverse collection enables sharing of searches among colleagues with discoverable metadata fields for searching within deposited searches. While the initial creation of the DMP and documentation took about six hours, the subsequent deposit of search strategies into the institutional data repository requires minimal effort (e.g., 5-10 minutes on average per deposit). The Dataverse collection also empowers librarians to retain intellectual ownership over search strategies as valuable stand-alone research outputs and raise the visibility of their labor. Overall, institutional data repositories provide specific benefits in facilitating compliance both with PRISMA-S guidance and with RDM best practices.


Assuntos
Armazenamento e Recuperação da Informação , Humanos , Armazenamento e Recuperação da Informação/métodos , Disseminação de Informação/métodos , Gerenciamento de Dados/métodos , Bibliotecas Médicas/organização & administração , Bibliotecários/estatística & dados numéricos
9.
Brief Bioinform ; 22(1): 30-44, 2021 01 18.
Artigo em Inglês | MEDLINE | ID: mdl-32496509

RESUMO

Thousands of new experimental datasets are becoming available every day; in many cases, they are produced within the scope of large cooperative efforts, involving a variety of laboratories spread all over the world, and typically open for public use. Although the potential collective amount of available information is huge, the effective combination of such public sources is hindered by data heterogeneity, as the datasets exhibit a wide variety of notations and formats, concerning both experimental values and metadata. Thus, data integration is becoming a fundamental activity, to be performed prior to data analysis and biological knowledge discovery, consisting of subsequent steps of data extraction, normalization, matching and enrichment; once applied to heterogeneous data sources, it builds multiple perspectives over the genome, leading to the identification of meaningful relationships that could not be perceived by using incompatible data formats. In this paper, we first describe a technological pipeline from data production to data integration; we then propose a taxonomy of genomic data players (based on the distinction between contributors, repository hosts, consortia, integrators and consumers) and apply the taxonomy to describe about 30 important players in genomic data management. We specifically focus on the integrator players and analyse the issues in solving the genomic data integration challenges, as well as evaluate the computational environments that they provide to follow up data integration by means of visualization and analysis tools.


Assuntos
Gerenciamento de Dados/métodos , Genoma Humano , Genômica/métodos , Humanos , Metadados
10.
Brief Bioinform ; 22(1): 45-54, 2021 01 18.
Artigo em Inglês | MEDLINE | ID: mdl-32533135

RESUMO

With advances in genomic sequencing technology, a large amount of data is publicly available for the research community to extract meaningful and reliable associations among risk genes and the mechanisms of disease. However, this exponential growth of data is spread in over thousand heterogeneous repositories, represented in multiple formats and with different levels of quality what hinders the differentiation of clinically valid relationships from those that are less well-sustained and that could lead to wrong diagnosis. This paper presents how conceptual models can play a key role to efficiently manage genomic data. These data must be accessible, informative and reliable enough to extract valuable knowledge in the context of the identification of evidence supporting the relationship between DNA variants and disease. The approach presented in this paper provides a solution that help researchers to organize, store and process information focusing only on the data that are relevant and minimizing the impact that the information overload has in clinical and research contexts. A case-study (epilepsy) is also presented, to demonstrate its application in a real context.


Assuntos
Gerenciamento de Dados/métodos , Genômica/métodos , Sistemas de Dados , Epilepsia/genética , Predisposição Genética para Doença , Humanos
11.
Brief Bioinform ; 22(1): 474-484, 2021 01 18.
Artigo em Inglês | MEDLINE | ID: mdl-31885044

RESUMO

BACKGROUND: With the increasing development of biotechnology and information technology, publicly available data in chemistry and biology are undergoing explosive growth. Such wealthy information in these resources needs to be extracted and then transformed to useful knowledge by various data mining methods. However, a main computational challenge is how to effectively represent or encode molecular objects under investigation such as chemicals, proteins, DNAs and even complicated interactions when data mining methods are employed. To further explore these complicated data, an integrated toolkit to represent different types of molecular objects and support various data mining algorithms is urgently needed. RESULTS: We developed a freely available R/CRAN package, called BioMedR, for molecular representations of chemicals, proteins, DNAs and pairwise samples of their interactions. The current version of BioMedR could calculate 293 molecular descriptors and 13 kinds of molecular fingerprints for small molecules, 9920 protein descriptors based on protein sequences and six types of generalized scale-based descriptors for proteochemometric modeling, more than 6000 DNA descriptors from nucleotide sequences and six types of interaction descriptors using three different combining strategies. Moreover, this package realized five similarity calculation methods and four powerful clustering algorithms as well as several useful auxiliary tools, which aims at building an integrated analysis pipeline for data acquisition, data checking, descriptor calculation and data modeling. CONCLUSION: BioMedR provides a comprehensive and uniform R package to link up different representations of molecular objects with each other and will benefit cheminformatics/bioinformatics and other biomedical users. It is available at: https://CRAN.R-project.org/package=BioMedR and https://github.com/wind22zhu/BioMedR/.


Assuntos
Biologia Computacional/métodos , Sistemas de Gerenciamento de Base de Dados , Gerenciamento de Dados/métodos , Bases de Dados de Compostos Químicos , Bases de Dados Genéticas , Humanos
12.
Brief Bioinform ; 22(5)2021 09 02.
Artigo em Inglês | MEDLINE | ID: mdl-33589928

RESUMO

This article describes some use case studies and self-assessments of FAIR status of de.NBI services to illustrate the challenges and requirements for the definition of the needs of adhering to the FAIR (findable, accessible, interoperable and reusable) data principles in a large distributed bioinformatics infrastructure. We address the challenge of heterogeneity of wet lab technologies, data, metadata, software, computational workflows and the levels of implementation and monitoring of FAIR principles within the different bioinformatics sub-disciplines joint in de.NBI. On the one hand, this broad service landscape and the excellent network of experts are a strong basis for the development of useful research data management plans. On the other hand, the large number of tools and techniques maintained by distributed teams renders FAIR compliance challenging.


Assuntos
Gerenciamento de Dados/métodos , Metadados , Redes Neurais de Computação , Proteômica/métodos , Software , Genoma Humano , Sequenciamento de Nucleotídeos em Larga Escala , Humanos , Cooperação Internacional , Fenótipo , Plantas/genética , Proteoma , Autoavaliação (Psicologia) , Fluxo de Trabalho
14.
Proc Natl Acad Sci U S A ; 117(9): 4571-4577, 2020 03 03.
Artigo em Inglês | MEDLINE | ID: mdl-32071251

RESUMO

Machine learning is proving invaluable across disciplines. However, its success is often limited by the quality and quantity of available data, while its adoption is limited by the level of trust afforded by given models. Human vs. machine performance is commonly compared empirically to decide whether a certain task should be performed by a computer or an expert. In reality, the optimal learning strategy may involve combining the complementary strengths of humans and machines. Here, we present expert-augmented machine learning (EAML), an automated method that guides the extraction of expert knowledge and its integration into machine-learned models. We used a large dataset of intensive-care patient data to derive 126 decision rules that predict hospital mortality. Using an online platform, we asked 15 clinicians to assess the relative risk of the subpopulation defined by each rule compared to the total sample. We compared the clinician-assessed risk to the empirical risk and found that, while clinicians agreed with the data in most cases, there were notable exceptions where they overestimated or underestimated the true risk. Studying the rules with greatest disagreement, we identified problems with the training data, including one miscoded variable and one hidden confounder. Filtering the rules based on the extent of disagreement between clinician-assessed risk and empirical risk, we improved performance on out-of-sample data and were able to train with less data. EAML provides a platform for automated creation of problem-specific priors, which help build robust and dependable machine-learning models in critical applications.


Assuntos
Sistemas Inteligentes , Aprendizado de Máquina/normas , Informática Médica/métodos , Gerenciamento de Dados/métodos , Sistemas de Gerenciamento de Base de Dados , Informática Médica/normas
15.
PLoS Biol ; 17(6): e3000343, 2019 06.
Artigo em Inglês | MEDLINE | ID: mdl-31220074

RESUMO

The zebrafish Danio rerio is a powerful model system to study the genetics of development and disease. However, maintenance of zebrafish husbandry records is both time intensive and laborious, and a standardized way to manage and track the large amount of unique lines in a given laboratory or centralized facility has not been embraced by the field. Here, we present FishNET, an intuitive, open-source, relational database for managing data and information related to zebrafish husbandry and maintenance. By creating a "virtual facility," FishNET enables users to remotely inspect the rooms, racks, tanks, and lines within a given facility. Importantly, FishNET scales from one laboratory to an entire facility with several laboratories to multiple facilities, generating a cohesive laboratory and community-based platform. Automated data entry eliminates confusion regarding line nomenclature and streamlines maintenance of individual lines, while flexible query forms allow researchers to retrieve database records based on user-defined criteria. FishNET also links associated embryonic and adult biological samples with data, such as genotyping results or confocal images, to enable robust and efficient colony management and storage of laboratory information. A shared calendar function with email notifications and automated reminders for line turnover, automated tank counts, and census reports promote communication with both end users and administrators. The expected benefits of FishNET are improved vivaria efficiency, increased quality control for experimental numbers, and flexible data reporting and retrieval. FishNET's easy, intuitive record management and open-source, end-user-modifiable architecture provides an efficient solution to real-time zebrafish colony management for users throughout a facility and institution and, in some cases, across entire research hubs.


Assuntos
Criação de Animais Domésticos/métodos , Peixe-Zebra , Criação de Animais Domésticos/normas , Animais , Gerenciamento de Dados/métodos , Sistemas de Gerenciamento de Base de Dados , Bases de Dados Factuais , Laboratórios , Software
16.
PLoS Biol ; 17(8): e3000384, 2019 08.
Artigo em Inglês | MEDLINE | ID: mdl-31404057

RESUMO

Citation metrics are widely used and misused. We have created a publicly available database of 100,000 top scientists that provides standardized information on citations, h-index, coauthorship-adjusted hm-index, citations to papers in different authorship positions, and a composite indicator. Separate data are shown for career-long and single-year impact. Metrics with and without self-citations and ratio of citations to citing papers are given. Scientists are classified into 22 scientific fields and 176 subfields. Field- and subfield-specific percentiles are also provided for all scientists who have published at least five papers. Career-long data are updated to end of 2017 and to end of 2018 for comparison.


Assuntos
Autoria/normas , Curadoria de Dados/métodos , Bases de Dados Factuais/normas , Bibliometria , Gerenciamento de Dados/métodos , Humanos , Fator de Impacto de Revistas , Publicações/tendências , Editoração/tendências , Padrões de Referência , Pesquisadores
17.
Anesth Analg ; 134(2): 380-388, 2022 02 01.
Artigo em Inglês | MEDLINE | ID: mdl-34673658

RESUMO

BACKGROUND: The retrospective analysis of electroencephalogram (EEG) signals acquired from patients under general anesthesia is crucial in understanding the patient's unconscious brain's state. However, the creation of such database is often tedious and cumbersome and involves human labor. Hence, we developed a Raspberry Pi-based system for archiving EEG signals recorded from patients under anesthesia in operating rooms (ORs) with minimal human involvement. METHODS: Using this system, we archived patient EEG signals from over 500 unique surgeries at the Emory University Orthopaedics and Spine Hospital, Atlanta, for about 18 months. For this, we developed a software package that runs on a Raspberry Pi and archives patient EEG signals from a SedLine Root EEG Monitor (Masimo) to a secure Health Insurance Portability and Accountability Act (HIPAA) compliant cloud storage. The OR number corresponding to each surgery was archived along with the EEG signal to facilitate retrospective EEG analysis. We retrospectively processed the archived EEG signals and performed signal quality checks. We also proposed a formula to compute the proportion of true EEG signal and calculated the corresponding statistics. Further, we curated and interleaved patient medical record information with the corresponding EEG signals. RESULTS: We retrospectively processed the EEG signals to demonstrate a statistically significant negative correlation between the relative alpha power (8-12 Hz) of the EEG signal captured under anesthesia and the patient's age. CONCLUSIONS: Our system is a standalone EEG archiver developed using low cost and readily available hardware. We demonstrated that one could create a large-scale EEG database with minimal human involvement. Moreover, we showed that the captured EEG signal is of good quality for retrospective analysis and combined the EEG signal with the patient medical records. This project's software has been released under an open-source license to enable others to use and contribute.


Assuntos
Curadoria de Dados/métodos , Eletroencefalografia/instrumentação , Eletroencefalografia/métodos , Monitorização Intraoperatória/instrumentação , Monitorização Intraoperatória/métodos , Adulto , Idoso , Idoso de 80 Anos ou mais , Gerenciamento de Dados/instrumentação , Gerenciamento de Dados/métodos , Feminino , Humanos , Masculino , Pessoa de Meia-Idade , Estudos Retrospectivos , Adulto Jovem
18.
Retina ; 42(1): 4-10, 2022 01 01.
Artigo em Inglês | MEDLINE | ID: mdl-34081638

RESUMO

PURPOSE: To review the current literature on the management of proliferative diabetic retinopathy (PDR) and the challenges in the real-world setting. METHODS: A review of the literature was performed on the therapeutic options for PDR, with a focus on the real-world data presented by the Pan-American Collaborative Retina Study Group. RESULTS: Data from clinical trials and previous literature have reported that intravitreal antivascular endothelial growth factor (anti-VEGF) therapy is noninferior to the gold standard panretinal photocoagulation for treating PDR. However, PDR recurs rapidly after cessation of anti-VEGF therapy. This is especially important in the context of the diabetic population that is prone to loss to follow-up. In a real-world, prospective study, patients with prior panretinal photocoagulation followed by anti-VEGF therapy had higher rates of sustained PDR regression relative to anti-VEGF therapy alone. CONCLUSION: Owing to its transient therapeutic effect, anti-VEGF therapy in patients with diabetes can present a risk of recurrent retinal neovascularization and progression of PDR if follow-up cannot be guaranteed. A combined paradigm with less aggressive, immediate panretinal photocoagulation followed by anti-VEGF therapy should be considered in this population.


Assuntos
Gerenciamento de Dados/métodos , Retinopatia Diabética/terapia , Gerenciamento Clínico , Retinopatia Diabética/epidemiologia , Humanos , América Latina/epidemiologia , Morbidade , Espanha/epidemiologia
19.
Biochemistry ; 60(38): 2902-2914, 2021 09 28.
Artigo em Inglês | MEDLINE | ID: mdl-34491035

RESUMO

Citrullination is an enzyme-catalyzed post-translational modification (PTM) that is essential for a host of biological processes, including gene regulation, programmed cell death, and organ development. While this PTM is required for normal cellular functions, aberrant citrullination is a hallmark of autoimmune disorders as well as cancer. Although aberrant citrullination is linked to human pathology, the exact role of citrullination in disease remains poorly characterized, in part because of the challenges associated with identifying the specific arginine residues that are citrullinated. Tandem mass spectrometry is the most precise method for uncovering sites of citrullination; however, due to the small mass shift (+0.984 Da) that results from citrullination, current database search algorithms commonly misannotate spectra, leading to a high number of false-positive assignments. To address this challenge, we developed an automated workflow to rigorously and rapidly mine proteomic data to unambiguously identify the sites of citrullination from complex peptide mixtures. The crux of this streamlined workflow is the ionFinder software program, which classifies citrullination sites with high confidence on the basis of the presence of diagnostic fragment ions. These diagnostic ions include the neutral loss of isocyanic acid, which is a dissociative event that is unique to citrulline residues. Using the ionFinder program, we have mapped the sites of autocitrullination on purified protein arginine deiminases (PAD1-4) and mapped the global citrullinome in a PAD2-overexpressing cell line. The ionFinder algorithm is a highly versatile, user-friendly, and open-source program that is agnostic to the type of instrument and mode of fragmentation that are used.


Assuntos
Citrulinação/fisiologia , Mineração de Dados/métodos , Proteômica/métodos , Algoritmos , Arginina/metabolismo , Citrulinação/genética , Citrulina/química , Citrulina/genética , Citrulina/metabolismo , Análise de Dados , Gerenciamento de Dados/métodos , Humanos , Peptídeos/metabolismo , Processamento de Proteína Pós-Traducional , Desiminases de Arginina em Proteínas/genética , Desiminases de Arginina em Proteínas/metabolismo , Espectrometria de Massas em Tandem/métodos
20.
Radiology ; 301(1): 115-122, 2021 10.
Artigo em Inglês | MEDLINE | ID: mdl-34342503

RESUMO

Background Patterns of metastasis in cancer are increasingly relevant to prognostication and treatment planning but have historically been documented by means of autopsy series. Purpose To show the feasibility of using natural language processing (NLP) to gather accurate data from radiology reports for assessing spatial and temporal patterns of metastatic spread in a large patient cohort. Materials and Methods In this retrospective longitudinal study, consecutive patients who underwent CT from July 2009 to April 2019 and whose CT reports followed a departmental structured template were included. Three radiologists manually curated a sample of 2219 reports for the presence or absence of metastases across 13 organs; these manually curated reports were used to develop three NLP models with an 80%-20% split for training and test sets. A separate random sample of 448 manually curated reports was used for validation. Model performance was measured by accuracy, precision, and recall for each organ. The best-performing NLP model was used to generate a final database of metastatic disease across all patients. For each cancer type, statistical descriptive reports were provided by analyzing the frequencies of metastatic disease at the report and patient levels. Results In 91 665 patients (mean age ± standard deviation, 61 years ± 15; 46 939 women), 387 359 reports were labeled. The best-performing NLP model achieved accuracies from 90% to 99% across all organs. Metastases were most frequently reported in abdominopelvic (23.6% of all reports) and thoracic (17.6%) nodes, followed by lungs (14.7%), liver (13.7%), and bones (9.9%). Metastatic disease tropism is distinct among common cancers, with the most common first site being bones in prostate and breast cancers and liver among pancreatic and colorectal cancers. Conclusion Natural language processing may be applied to cancer patients' CT reports to generate a large database of metastatic phenotypes. Such a database could be combined with genomic studies and used to explore prognostic imaging phenotypes with relevance to treatment planning. © RSNA, 2021 Online supplemental material is available for this article.


Assuntos
Gerenciamento de Dados/métodos , Bases de Dados Factuais/estatística & dados numéricos , Registros Eletrônicos de Saúde , Processamento de Linguagem Natural , Neoplasias/epidemiologia , Tomografia Computadorizada por Raios X/métodos , Estudos de Viabilidade , Feminino , Humanos , Estudos Longitudinais , Masculino , Pessoa de Meia-Idade , Metástase Neoplásica , Reprodutibilidade dos Testes , Estudos Retrospectivos
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA