Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 67
Filtrar
1.
J Neurosci ; 44(38)2024 Sep 18.
Artigo em Inglês | MEDLINE | ID: mdl-39293939

RESUMO

Neuroscience research has evolved to generate increasingly large and complex experimental data sets, and advanced data science tools are taking on central roles in neuroscience research. Neurodata Without Borders (NWB), a standard language for neurophysiology data, has recently emerged as a powerful solution for data management, analysis, and sharing. We here discuss our labs' efforts to implement NWB data science pipelines. We describe general principles and specific use cases that illustrate successes, challenges, and non-trivial decisions in software engineering. We hope that our experience can provide guidance for the neuroscience community and help bridge the gap between experimental neuroscience and data science. Key takeaways from this article are that (1) standardization with NWB requires non-trivial design choices; (2) the general practice of standardization in the lab promotes data awareness and literacy, and improves transparency, rigor, and reproducibility in our science; (3) we offer several feature suggestions to ease the extensibility, publishing/sharing, and usability for NWB standard and users of NWB data.


Assuntos
Neurociências , Neurociências/normas , Neurociências/métodos , Humanos , Software/normas , Disseminação de Informação/métodos , Ciência de Dados/métodos , Animais
2.
Anal Bioanal Chem ; 416(9): 2125-2136, 2024 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-38300263

RESUMO

This trend article provides an overview of recent advancements in Non-Target Screening (NTS) for water quality assessment, focusing on new methods in data evaluation, qualification, quantification, and quality assurance (QA/QC). It highlights the evolution in NTS data processing, where open-source platforms address challenges in result comparability and data complexity. Advanced chemometrics and machine learning (ML) are pivotal for trend identification and correlation analysis, with a growing emphasis on automated workflows and robust classification models. The article also discusses the rigorous QA/QC measures essential in NTS, such as internal standards, batch effect monitoring, and matrix effect assessment. It examines the progress in quantitative NTS (qNTS), noting advancements in ionization efficiency-based quantification and predictive modeling despite challenges in sample variability and analytical standards. Selected studies illustrate NTS's role in water analysis, combining high-resolution mass spectrometry with chromatographic techniques for enhanced chemical exposure assessment. The article addresses chemical identification and prioritization challenges, highlighting the integration of database searches and computational tools for efficiency. Finally, the article outlines the future research needs in NTS, including establishing comprehensive guidelines, improving QA/QC measures, and reporting results. It underscores the potential to integrate multivariate chemometrics, AI/ML tools, and multi-way methods into NTS workflows and combine various data sources to understand ecosystem health and protection comprehensively.

3.
J Med Internet Res ; 26: e56614, 2024 May 31.
Artigo em Inglês | MEDLINE | ID: mdl-38819879

RESUMO

BACKGROUND: Efficient data exchange and health care interoperability are impeded by medical records often being in nonstandardized or unstructured natural language format. Advanced language models, such as large language models (LLMs), may help overcome current challenges in information exchange. OBJECTIVE: This study aims to evaluate the capability of LLMs in transforming and transferring health care data to support interoperability. METHODS: Using data from the Medical Information Mart for Intensive Care III and UK Biobank, the study conducted 3 experiments. Experiment 1 assessed the accuracy of transforming structured laboratory results into unstructured format. Experiment 2 explored the conversion of diagnostic codes between the coding frameworks of the ICD-9-CM (International Classification of Diseases, Ninth Revision, Clinical Modification), and Systematized Nomenclature of Medicine Clinical Terms (SNOMED-CT) using a traditional mapping table and a text-based approach facilitated by the LLM ChatGPT. Experiment 3 focused on extracting targeted information from unstructured records that included comprehensive clinical information (discharge notes). RESULTS: The text-based approach showed a high conversion accuracy in transforming laboratory results (experiment 1) and an enhanced consistency in diagnostic code conversion, particularly for frequently used diagnostic names, compared with the traditional mapping approach (experiment 2). In experiment 3, the LLM showed a positive predictive value of 87.2% in extracting generic drug names. CONCLUSIONS: This study highlighted the potential role of LLMs in significantly improving health care data interoperability, demonstrated by their high accuracy and efficiency in data transformation and exchange. The LLMs hold vast potential for enhancing medical data exchange without complex standardization for medical terms and data structure.


Assuntos
Troca de Informação em Saúde , Humanos , Troca de Informação em Saúde/normas , Interoperabilidade da Informação em Saúde , Registros Eletrônicos de Saúde , Processamento de Linguagem Natural , Systematized Nomenclature of Medicine
4.
Alzheimers Dement ; 2024 Aug 14.
Artigo em Inglês | MEDLINE | ID: mdl-39140601

RESUMO

The goal of the Biostatistics Core of the Alzheimer's Disease Neuroimaging Initiative (ADNI) has been to ensure that sound study designs and statistical methods are used to meet the overall goals of ADNI. We have supported the creation of a well-validated and well-curated longitudinal database of clinical and biomarker information on ADNI participants and helped to make this accessible and usable for researchers. We have developed a statistical methodology for characterizing the trajectories of clinical and biomarker change for ADNI participants across the spectrum from cognitively normal to dementia, including multivariate patterns and evidence for heterogeneity in cognitive aging. We have applied these methods and adapted them to improve clinical trial design. ADNI-4 will offer us a chance to help extend these efforts to a more diverse cohort with an even richer panel of biomarker data to support better knowledge of and treatment for Alzheimer's disease and related dementias. HIGHLIGHTS: The Alzheimer's Disease Neuroimaging Initiative (ADNI) Biostatistics Core provides study design and analytic support to ADNI investigators. Core members develop and apply novel statistical methodology to work with ADNI data and support clinical trial design. The Core contributes to the standardization, validation, and harmonization of biomarker data. The Core serves as a resource to the wider research community to address questions related to the data and study as a whole.

5.
J Med Internet Res ; 25: e46165, 2023 07 20.
Artigo em Inglês | MEDLINE | ID: mdl-37471130

RESUMO

BACKGROUND: Mood disorder has emerged as a serious concern for public health; in particular, bipolar disorder has a less favorable prognosis than depression. Although prompt recognition of depression conversion to bipolar disorder is needed, early prediction is challenging due to overlapping symptoms. Recently, there have been attempts to develop a prediction model by using federated learning. Federated learning in medical fields is a method for training multi-institutional machine learning models without patient-level data sharing. OBJECTIVE: This study aims to develop and validate a federated, differentially private multi-institutional bipolar transition prediction model. METHODS: This retrospective study enrolled patients diagnosed with the first depressive episode at 5 tertiary hospitals in South Korea. We developed models for predicting bipolar transition by using data from 17,631 patients in 4 institutions. Further, we used data from 4541 patients for external validation from 1 institution. We created standardized pipelines to extract large-scale clinical features from the 4 institutions without any code modification. Moreover, we performed feature selection in a federated environment for computational efficiency and applied differential privacy to gradient updates. Finally, we compared the federated and the 4 local models developed with each hospital's data on internal and external validation data sets. RESULTS: In the internal data set, 279 out of 17,631 patients showed bipolar disorder transition. In the external data set, 39 out of 4541 patients showed bipolar disorder transition. The average performance of the federated model in the internal test (area under the curve [AUC] 0.726) and external validation (AUC 0.719) data sets was higher than that of the other locally developed models (AUC 0.642-0.707 and AUC 0.642-0.699, respectively). In the federated model, classifications were driven by several predictors such as the Charlson index (low scores were associated with bipolar transition, which may be due to younger age), severe depression, anxiolytics, young age, and visiting months (the bipolar transition was associated with seasonality, especially during the spring and summer months). CONCLUSIONS: We developed and validated a differentially private federated model by using distributed multi-institutional psychiatric data with standardized pipelines in a real-world environment. The federated model performed better than models using local data only.


Assuntos
Transtorno Bipolar , Aprendizado de Máquina , Privacidade , Humanos , Transtorno Bipolar/diagnóstico , Depressão/diagnóstico , Transtornos do Humor , Estudos Retrospectivos
6.
BMC Bioinformatics ; 23(Suppl 12): 386, 2022 Sep 23.
Artigo em Inglês | MEDLINE | ID: mdl-36151511

RESUMO

BACKGROUND: Public Data Commons (PDC) have been highlighted in the scientific literature for their capacity to collect and harmonize big data. On the other hand, local data commons (LDC), located within an institution or organization, have been underrepresented in the scientific literature, even though they are a critical part of research infrastructure. Being closest to the sources of data, LDCs provide the ability to collect and maintain the most up-to-date, high-quality data within an organization, closest to the sources of the data. As a data provider, LDCs have many challenges in both collecting and standardizing data, moreover, as a consumer of PDC, they face problems of data harmonization stemming from the monolithic harmonization pipeline designs commonly adapted by many PDCs. Unfortunately, existing guidelines and resources for building and maintaining data commons exclusively focus on PDC and provide very little information on LDC. RESULTS: This article focuses on four important observations. First, there are three different types of LDC service models that are defined based on their roles and requirements. These can be used as guidelines for building new LDC or enhancing the services of existing LDC. Second, the seven core services of LDC are discussed, including cohort identification and facilitation of genomic sequencing, the management of molecular reports and associated infrastructure, quality control, data harmonization, data integration, data sharing, and data access control. Third, instead of commonly developed monolithic systems, we propose a new data sharing method for data harmonization that combines both divide-and-conquer and bottom-up approaches. Finally, an end-to-end LDC implementation is introduced with real-world examples. CONCLUSIONS: Although LDCs are an optimal place to identify and address data quality issues, they have traditionally been relegated to the role of passive data provider for much larger PDC. Indeed, many LDCs limit their functions to only conducting routine data storage and transmission tasks due to a lack of information on how to design, develop, and improve their services using limited resources. We hope that this work will be the first small step in raising awareness among the LDCs of their expanded utility and to publicize to a wider audience the importance of LDC.


Assuntos
Big Data , Disseminação de Informação , Países em Desenvolvimento , Humanos
7.
Hum Mutat ; 43(6): 717-733, 2022 06.
Artigo em Inglês | MEDLINE | ID: mdl-35178824

RESUMO

Rare disease patients are more likely to receive a rapid molecular diagnosis nowadays thanks to the wide adoption of next-generation sequencing. However, many cases remain undiagnosed even after exome or genome analysis, because the methods used missed the molecular cause in a known gene, or a novel causative gene could not be identified and/or confirmed. To address these challenges, the RD-Connect Genome-Phenome Analysis Platform (GPAP) facilitates the collation, discovery, sharing, and analysis of standardized genome-phenome data within a collaborative environment. Authorized clinicians and researchers submit pseudonymised phenotypic profiles encoded using the Human Phenotype Ontology, and raw genomic data which is processed through a standardized pipeline. After an optional embargo period, the data are shared with other platform users, with the objective that similar cases in the system and queries from peers may help diagnose the case. Additionally, the platform enables bidirectional discovery of similar cases in other databases from the Matchmaker Exchange network. To facilitate genome-phenome analysis and interpretation by clinical researchers, the RD-Connect GPAP provides a powerful user-friendly interface and leverages tens of information sources. As a result, the resource has already helped diagnose hundreds of rare disease patients and discover new disease causing genes.


Assuntos
Genômica , Doenças Raras , Exoma , Estudos de Associação Genética , Genômica/métodos , Humanos , Fenótipo , Doenças Raras/diagnóstico , Doenças Raras/genética
8.
Dermatology ; 238(1): 44-52, 2022.
Artigo em Inglês | MEDLINE | ID: mdl-33735862

RESUMO

BACKGROUND: The Observational Health Data Sciences and Informatics (OHDSI) network enables access to billions of deidentified, standardized health records and built-in analytics software for observational health research, with numerous potential applications to dermatology. While the use of the OHDSI has increased steadily over the past several years, review of the literature reveals few studies utilizing OHDSI in dermatology. To our knowledge, the University of Colorado School of Medicine is unique in its use of OHDSI for dermatology big data research. SUMMARY: A PubMed search was conducted in August 2020, followed by a literature review, with 24 of the 72 screened articles selected for inclusion. In this review, we discuss the ways OHDSI has been used to compile and analyze data, improve prediction and estimation capabilities, and inform treatment guidelines across specialties. We also discuss the potential for OHDSI in dermatology - specifically, ways that it could reveal adherence to available guidelines, establish standardized protocols, and ensure health equity. Key Messages: OHDSI has demonstrated broad utility in medicine. Adoption of OHDSI by the field of dermatology would facilitate big data research, allow for examination of current prescribing and treatment patterns without clear best practice guidelines, improve the dermatologic knowledge base and, by extension, improve patient outcomes.


Assuntos
Pesquisa Biomédica/tendências , Ciência de Dados , Dermatologia/tendências , Informática Médica , Big Data , Humanos
9.
J Med Internet Res ; 22(10): e19879, 2020 10 07.
Artigo em Inglês | MEDLINE | ID: mdl-33026356

RESUMO

BACKGROUND: The introduction of next-generation sequencing (NGS) into molecular cancer diagnostics has led to an increase in the data available for the identification and evaluation of driver mutations and for defining personalized cancer treatment regimens. The meaningful combination of omics data, ie, pathogenic gene variants and alterations with other patient data, to understand the full picture of malignancy has been challenging. OBJECTIVE: This study describes the implementation of a system capable of processing, analyzing, and subsequently combining NGS data with other clinical patient data for analysis within and across institutions. METHODS: On the basis of the already existing NGS analysis workflows for the identification of malignant gene variants at the Institute of Pathology of the University Hospital Erlangen, we defined basic requirements on an NGS processing and analysis pipeline and implemented a pipeline based on the GEMINI (GEnome MINIng) open source genetic variation database. For the purpose of validation, this pipeline was applied to data from the 1000 Genomes Project and subsequently to NGS data derived from 206 patients of a local hospital. We further integrated the pipeline into existing structures of data integration centers at the University Hospital Erlangen and combined NGS data with local nongenomic patient-derived data available in Fast Healthcare Interoperability Resources format. RESULTS: Using data from the 1000 Genomes Project and from the patient cohort as input, the implemented system produced the same results as already established methodologies. Further, it satisfied all our identified requirements and was successfully integrated into the existing infrastructure. Finally, we showed in an exemplary analysis how the data could be quickly loaded into and analyzed in KETOS, a web-based analysis platform for statistical analysis and clinical decision support. CONCLUSIONS: This study demonstrates that the GEMINI open source database can be augmented to create an NGS analysis pipeline. The pipeline generates high-quality results consistent with the already established workflows for gene variant annotation and pathological evaluation. We further demonstrate how NGS-derived genomic and other clinical data can be combined for further statistical analysis, thereby providing for data integration using standardized vocabularies and methods. Finally, we demonstrate the feasibility of the pipeline integration into hospital workflows by providing an exemplary integration into the data integration center infrastructure, which is currently being established across Germany.


Assuntos
Sistemas de Apoio a Decisões Clínicas/normas , Atenção à Saúde/métodos , Genômica/métodos , Interoperabilidade da Informação em Saúde/normas , Internet/normas , Aprendizado de Máquina/normas , Humanos
10.
J Biomed Inform ; 91: 103119, 2019 03.
Artigo em Inglês | MEDLINE | ID: mdl-30738946

RESUMO

OBJECTIVE: Supplementing the Spontaneous Reporting System (SRS) with Electronic Health Record (EHR) data for adverse drug reaction detection could augment sample size, increase population heterogeneity and cross-validate results for pharmacovigilance research. The difference in the underlying data structures and terminologies between SRS and EHR data presents challenges when attempting to integrate the two into a single database. The Observational Health Data Sciences and Informatics (OHDSI) collaboration provides a Common Data Model (CDM) for organizing and standardizing EHR data to support large-scale observational studies. The objective of the study is to develop and evaluate an informatics platform known as ADEpedia-on-OHDSI, where spontaneous reporting data from FDA's Adverse Event Reporting System (FAERS) is converted into the OHDSI CDM format towards building a next generation pharmacovigilance signal detection platform. METHODS: An extraction, transformation and loading (ETL) tool was designed, developed, and implemented to convert FAERS data into the OHDSI CDM format. A comprehensive evaluation, including overall ETL evaluation, mapping quality evaluation of drug names to RxNorm, and an evaluation of transformation and imputation quality, was then performed to assess the mapping accuracy and information loss using the FAERS data collected between 2012 and 2017. Previously published findings related to vascular safety profile of triptans were validated using ADEpedia-on-OHDSI in pharmacovigilance research. For the triptan-related vascular event detection, signals were detected by Reporting Odds Ratio (ROR) in high-level group terms (HLGT) level, high-level terms (HLT) level and preferred term (PT) level using the original FAERS data and CDM-based FAERS respectively. In addition, six standardized MedDRA queries (SMQs) related to vascular events were applied. RESULTS: A total of 4,619,362 adverse event cases were loaded into 8 tables in the OHDSI CDM. For drug name mapping, 93.9% records and 47.0% unique names were matched with RxNorm codes. Mapping accuracy of drug names was 96% based on a manual verification of randomly sampled 500 unique mappings. Information loss evaluation showed that more than 93% of the data is loaded into the OHDSI CDM for most fields, with the exception of drug route data (66%). The replication study detected 5, 18, 47 and 6, 18, 50 triptan-related vascular event signals in MedDRA HLGT level, HLT level, and PT level for the original FAERS data and CDM-based FAERS respectively. The signal detection scores of six standardized MedDRA queries (SMQs) of vascular events in the raw data study were found to be lower than those scores in the CDM study. CONCLUSION: The outcome of this work would facilitate seamless integration and combined analyses of both SRS and EHR data for pharmacovigilance in ADEpedia-on-OHDSI, our platform for next generation pharmacovigilance.


Assuntos
Sistemas de Notificação de Reações Adversas a Medicamentos , Simulação por Computador , Farmacovigilância , Humanos , Estados Unidos
11.
Neurocrit Care ; 30(Suppl 1): 87-101, 2019 06.
Artigo em Inglês | MEDLINE | ID: mdl-31102238

RESUMO

INTRODUCTION: Variability in usage and definition of data characteristics in previous cohort studies on unruptured intracranial aneurysms (UIA) complicated pooling and proper interpretation of these data. The aim of the National Institute of Health/National Institute of Neurological Disorders and Stroke UIA and Subarachnoid Hemorrhage (SAH) Common Data Elements (CDE) Project was to provide a common structure for data collection in future research on UIA and SAH. METHODS: This paper describes the development and summarization of the recommendations of the working groups (WGs) on UIAs, which consisted of an international and multidisciplinary panel of cerebrovascular specialists on research and treatment of UIAs. Consensus recommendations were developed by review of previously published CDEs for other neurological diseases and the literature on UIAs. Recommendations for CDEs were classified by priority into 'Core,' 'Supplemental-Highly Recommended,' 'Supplemental,' and 'Exploratory.' RESULTS: Ninety-one CDEs were compiled; 69 were newly created and 22 were existing CDEs. The CDEs were assigned to eight subcategories and were classified as Core (8), Supplemental-Highly Recommended (23), Supplemental (25), and Exploratory (35) elements. Additionally, the WG developed and agreed on a classification for aneurysm morphology. CONCLUSION: The proposed CDEs have been distilled from a broad pool of characteristics, measures, or outcomes. The usage of these CDEs will facilitate pooling of data from cohort studies or clinical trials on patients with UIAs.


Assuntos
Elementos de Dados Comuns , Aneurisma Intracraniano , Pesquisa Biomédica , Ensaios Clínicos como Assunto , Estudos de Coortes , Humanos , National Institute of Neurological Disorders and Stroke (USA) , National Library of Medicine (U.S.) , Estados Unidos
12.
Neurocrit Care ; 30(Suppl 1): 60-78, 2019 06.
Artigo em Inglês | MEDLINE | ID: mdl-31115823

RESUMO

INTRODUCTION: Lack of homogeneous definitions for imaging data and consensus on their relevance in the setting of subarachnoid hemorrhage and unruptured intracranial aneurysms lead to a difficulty of data pooling and lack of robust data. The aim of the National Institute of Health/National Institute of Neurological Disorders and Stroke, Unruptured Intracranial Aneurysm (UIA) and Subarachnoid Hemorrhage (SAH) Common Data Elements (CDE) Project was to standardize data elements to ultimately facilitate data pooling and establish a more robust data quality in future neurovascular research on UIA and SAH. METHODS: For the subcommittee 'Radiological imaging of SAH,' international cerebrovascular specialists with imaging expertise in the setting of SAH were selected by the steering committee. CDEs were developed after reviewing the literature on neuroradiology and already existing CDEs for other neurological diseases. For prioritization, the CDEs were classified into 'Core,' 'Supplemental-Highly Recommended,' 'Supplemental' and 'Exploratory.' RESULTS: The subcommittee compiled 136 CDEs, 100 out of which were derived from previously established CDEs on ischemic stroke and 36 were newly created. The CDEs were assigned to four main categories (several CDEs were assigned to more than one category): 'Parenchymal imaging' with 42 CDEs, 'Angiography' with 49 CDEs, 'Perfusion imaging' with 20 CDEs, and 'Transcranial doppler' with 55 CDEs. The CDEs were classified into core, supplemental highly recommended, supplemental and exploratory elements. The core CDEs were imaging modality, imaging modality type, imaging modality vessel, angiography type, vessel angiography arterial anatomic site and imaging vessel angiography arterial result. CONCLUSIONS: The CDEs were established based on the current literature and consensus across cerebrovascular specialists. The use of these CDEs will facilitate standardization and aggregation of imaging data in the setting of SAH. However, the CDEs may require reevaluation and periodic adjustment based on current research and improved imaging quality and novel modalities.


Assuntos
Aneurisma Roto/diagnóstico por imagem , Elementos de Dados Comuns , Aneurisma Intracraniano/diagnóstico por imagem , Hemorragia Subaracnóidea/diagnóstico por imagem , Angiografia Digital , Pesquisa Biomédica , Angiografia Cerebral , Angiografia por Tomografia Computadorizada , Humanos , Angiografia por Ressonância Magnética , Imageamento por Ressonância Magnética , National Institute of Neurological Disorders and Stroke (USA) , National Library of Medicine (U.S.) , Imagem de Perfusão , Tomografia Computadorizada por Raios X , Ultrassonografia Doppler Transcraniana , Estados Unidos
13.
Zhongguo Zhong Yao Za Zhi ; 43(9): 1864-1870, 2018 May.
Artigo em Zh | MEDLINE | ID: mdl-29902898

RESUMO

To study the effect of different data standardization methods on the spectrum-effect relationship for anticoagulant effect of Trichosanthis Fructus dropping pills. The spectrum-effect relationship was studied by using grey correlation degree method between three doses of Trichosanthis Fructus dropping pills and prothrombintime (PT) in mice. The effect of 10 data standardization methods, namely minimization method, maximum method, data extreme difference method, standard deviation standardization method, initialization transformation method, mean transformation method, ratio of each chromatographic peak area to the total peak area, ratio of each chromatographic peak area to the common peak area, logarithmic standardization method and tangent normalization method on the spectrum-effect relationship between Trichosanthis Fructus dropping pills and PT in mice was evaluated by using relative correlation degree as the index. The results of spectrum-effect relationship can be expressed by the minimization method, the data extreme difference method, the standard deviation standardization method, the initialization method and the mean transformation method, with highest relative correlation degree by the mean transformation method. As compared with the mean transformation method, there were significant differences between the high dose group and the medium dose group in the minimization method and the data extreme difference method (P<0.01), while the minimization method in the low dose group showed statistical significance (P<0.05). The standard deviation standardization method, initialization method and the mean transformation method can be used to study the spectrum-effect relationship for the anticoagulation of Trichosanthis Fructus dropping pills.


Assuntos
Frutas , Animais , Anticoagulantes , Camundongos
14.
Adv Exp Med Biol ; 947: 303-324, 2017.
Artigo em Inglês | MEDLINE | ID: mdl-28168672

RESUMO

The particular properties of nanomaterials have led to their rapidly increasing use in diverse fields of application. However, safety assessment is not keeping pace and there are still gaps in the understanding of their hazards. Computational models predicting nanotoxicity, such as (quantitative) structure-activity relationships ((Q)SARs), can contribute to safety evaluation, in line with general efforts to apply alternative methods in chemical risk assessment. Their development is highly dependent on the availability of reliable and high quality experimental data, both regarding the compounds' properties as well as the measured toxic effects. In particular, "nano-QSARs" should take the nano-specific characteristics into account. The information compiled needs to be well organized, quality controlled and standardized. Integrating the data in an overarching, structured data collection aims to (a) organize the data in a way to support modelling, (b) make (meta)data necessary for modelling available, and (c) add value by making a comparison between data from different sources possible.Based on the available data, specific descriptors can be derived to parameterize the nanomaterial-specific structure and physico-chemical properties appropriately. Furthermore, the interactions between nanoparticles and biological systems as well as small molecules, which can lead to modifications of the structure of the active nanoparticles, need to be described and taken into account in the development of models to predict the biological activity and toxicity of nanoparticles. The EU NanoPUZZLES project was part of a global cooperative effort to advance data availability and modelling approaches supporting the characterization and evaluation of nanomaterials.


Assuntos
Nanopartículas/efeitos adversos , Nanopartículas/química , Simulação por Computador , Humanos , Nanoestruturas/efeitos adversos , Nanoestruturas/química , Relação Quantitativa Estrutura-Atividade , Medição de Risco
15.
Future Oncol ; 12(1): 119-36, 2016 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-26674745

RESUMO

The advances in diagnostic and treatment technology are responsible for a remarkable transformation in the internal medicine concept with the establishment of a new idea of personalized medicine. Inter- and intra-patient tumor heterogeneity and the clinical outcome and/or treatment's toxicity's complexity, justify the effort to develop predictive models from decision support systems. However, the number of evaluated variables coming from multiple disciplines: oncology, computer science, bioinformatics, statistics, genomics, imaging, among others could be very large thus making traditional statistical analysis difficult to exploit. Automated data-mining processes and machine learning approaches can be a solution to organize the massive amount of data, trying to unravel important interaction. The purpose of this paper is to describe the strategy to collect and analyze data properly for decision support and introduce the concept of an 'umbrella protocol' within the framework of 'rapid learning healthcare'.


Assuntos
Coleta de Dados , Mineração de Dados , Medicina de Precisão , Neoplasias Retais/epidemiologia , Humanos , Internet , Neoplasias Retais/tratamento farmacológico , Neoplasias Retais/patologia , Software
16.
Biochim Biophys Acta ; 1844(1 Pt A): 82-7, 2014 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-23524294

RESUMO

Significant advances have been made over the past ten years to standardize the data emerging from the proteomic workflows adopted by laboratories all over the world. Differences in workflows, instrumentation, analysis software and reporting methods initially resulted in very disparate data being generated by many of these research groups, making data storage and comparison challenging. As the data standards proposed by the HUPO-PSI have increasingly been adopted, and tools and databases implementing these data formats have become more readily available, data generated by these complex experimental procedures is now becoming easier to manipulate, to visualize and to analyse. Public domain databases now exist to collate the information generated by experimentalists and to make the generation of specific protein expression maps, and monitoring of changes in protein expression levels in response to external stimuli a real possibility. This article is part of a Special Issue entitled: Computational Proteomics in the Post-Identification Era. Guest Editors: Martin Eisenacher and Christian Stephan.


Assuntos
Interpretação Estatística de Dados , Proteômica , Bases de Dados de Proteínas
17.
Alzheimers Dement ; 11(10): 1212-21, 2015 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-25676387

RESUMO

INTRODUCTION: Data obtained in completed Alzheimer's disease (AD) clinical trials can inform decision making for future trials. Recognizing the importance of sharing these data, the Coalition Against Major Diseases created an Online Data Repository for AD (CODR-AD) with the aim of supporting accelerated drug development. The aim of this study was to build an open access, standardized database from control arm data collected across many clinical trials. METHODS: Comprehensive AD-specific data standards were developed to enable the pooling of data from different sources. Nine member organizations contributed patient-level data from 24 clinical trials of AD treatments. RESULTS: CODR-AD consists of control arm pooled and standardized data from 24 trials currently numbered at 6500 subjects; Alzheimer's Disease Assessment Scale-cognitive subscale 11 is the main outcome and specific covariates are also included. DISCUSSION: CODR-AD represents a unique integrated standardized clinical trials database available to qualified researchers. The pooling of data across studies facilitates a more comprehensive understanding of disease heterogeneity.


Assuntos
Doença de Alzheimer/tratamento farmacológico , Ensaios Clínicos como Assunto , Bases de Dados Factuais , Cognição , Tomada de Decisões , Humanos , Internet , Padrões de Referência , Estatística como Assunto
18.
J Sports Sci Med ; 13(2): 379-86, 2014 May.
Artigo em Inglês | MEDLINE | ID: mdl-24790493

RESUMO

Accelerometers are predominantly used to objectively measure the entire range of activity intensities - sedentary behaviour (SED), light physical activity (LPA) and moderate to vigorous physical activity (MVPA). However, studies consistently report results without accounting for systematic accelerometer wear-time variation (within and between participants), jeopardizing the validity of these results. This study describes the development of a standardization methodology to understand and minimize measurement bias due to wear-time variation. Accelerometry is generally conducted over seven consecutive days, with participants' data being commonly considered 'valid' only if wear-time is at least 10 hours/day. However, even within 'valid' data, there could be systematic wear-time variation. To explore this variation, accelerometer data of Smart Cities, Healthy Kids study (www.smartcitieshealthykids.com) were analyzed descriptively and with repeated measures multivariate analysis of variance (MANOVA). Subsequently, a standardization method was developed, where case-specific observed wear-time is controlled to an analyst specified time period. Next, case-specific accelerometer data are interpolated to this controlled wear-time to produce standardized variables. To understand discrepancies owing to wear-time variation, all analyses were conducted pre- and post-standardization. Descriptive analyses revealed systematic wear-time variation, both between and within participants. Pre- and post-standardized descriptive analyses of SED, LPA and MVPA revealed a persistent and often significant trend of wear-time's influence on activity. SED was consistently higher on weekdays before standardization; however, this trend was reversed post-standardization. Even though MVPA was significantly higher on weekdays both pre- and post-standardization, the magnitude of this difference decreased post-standardization. Multivariable analyses with standardized SED, LPA and MVPA as outcome variables yielded more stable results with narrower confidence intervals and smaller standard errors. Standardization of accelerometer data is effective in not only minimizing measurement bias due to systematic wear-time variation, but also to provide a uniform platform to compare results within and between populations and studies. Key pointsSystematic variation in accelerometer wear-time both, within and between participants results in measurement bias.Standardization of data after controlling for wear-time produces stable outcome variables.Descriptive and multivariate analyses conducted with standardized outcome variables minimize measurement bias.

19.
Stud Health Technol Inform ; 310: 48-52, 2024 Jan 25.
Artigo em Inglês | MEDLINE | ID: mdl-38269763

RESUMO

Observational Medical Outcome Partners - Common Data Model (OMOP-CDM) is an international standard model for standardizing electronic medical record data. However, unstructured data such as medical image data which is beyond the scope of standardization by the current OMOP-CDM is difficult to be used in multi-institutional collaborative research. Therefore, we developed the Radiology-CDM (R-CDM) which standardizes medical imaging data. As a proof of concept, 737,500 Optical Coherence Tomography (OCT) data from two tertiary hospitals in South Korea is standardized in the form of R-CDM. The relationship between chronic disease and retinal thickness was analyzed by using the R-CDM. Central macular thickness and retinal nerve fiber layer (RNFL) thickness were significantly thinner in the patients with hypertension compared to the control cohort. It is meaningful in that multi-institutional collaborative research using medical image data and clinical data simultaneously can be conducted very efficiently.


Assuntos
Face , Radiologia , Humanos , Radiografia , Retina/diagnóstico por imagem , Registros Eletrônicos de Saúde
20.
Stud Health Technol Inform ; 316: 354-355, 2024 Aug 22.
Artigo em Inglês | MEDLINE | ID: mdl-39176748

RESUMO

The growing number of genes identified in relation to epilepsy represents a major breakthrough in diagnosis and treatment, but experts face the challenge of efficiently accessing and consolidating the vast amount of genetic data available. Therefore, we present the process of transforming data from different sources and formats into an Entity-Attribute-Value (EAV) model database. Combined with the use of standard coding systems, this approach will provide a scalable and adaptable database to present the data in a comprehensive way to experts via a dashboard.


Assuntos
Epilepsia , Epilepsia/genética , Epilepsia/diagnóstico , Epilepsia/tratamento farmacológico , Humanos , Bases de Dados Genéticas
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA