Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 79
Filtrar
1.
Artigo em Alemão | MEDLINE | ID: mdl-38753021

RESUMO

The digital health progress hubs pilot the extensibility of the concepts and solutions of the Medical Informatics Initiative to improve regional healthcare and research. The six funded projects address different diseases, areas in regional healthcare, and methods of cross-institutional data linking and use. Despite the diversity of the scenarios and regional conditions, the technical, regulatory, and organizational challenges and barriers that the progress hubs encounter in the actual implementation of the solutions are often similar. This results in some common approaches to solutions, but also in political demands that go beyond the Health Data Utilization Act, which is considered a welcome improvement by the progress hubs.In this article, we present the digital progress hubs and discuss achievements, challenges, and approaches to solutions that enable the shared use of data from university hospitals and non-academic institutions in the healthcare system and can make a sustainable contribution to improving medical care and research.

2.
Artigo em Alemão | MEDLINE | ID: mdl-38750239

RESUMO

Health data are extremely important in today's data-driven world. Through automation, healthcare processes can be optimized, and clinical decisions can be supported. For any reuse of data, the quality, validity, and trustworthiness of data are essential, and it is the only way to guarantee that data can be reused sensibly. Specific requirements for the description and coding of reusable data are defined in the FAIR guiding principles for data stewardship. Various national research associations and infrastructure projects in the German healthcare sector have already clearly positioned themselves on the FAIR principles: both the infrastructures of the Medical Informatics Initiative and the University Medicine Network operate explicitly on the basis of the FAIR principles, as do the National Research Data Infrastructure for Personal Health Data and the German Center for Diabetes Research.To ensure that a resource complies with the FAIR principles, the degree of FAIRness should first be determined (so-called FAIR assessment), followed by the prioritization for improvement steps (so-called FAIRification). Since 2016, a set of tools and guidelines have been developed for both steps, based on the different, domain-specific interpretations of the FAIR principles.Neighboring European countries have also invested in the development of a national framework for semantic interoperability in the context of the FAIR (Findable, Accessible, Interoperable, Reusable) principles. Concepts for comprehensive data enrichment were developed to simplify data analysis, for example, in the European Health Data Space or via the Observational Health Data Sciences and Informatics network. With the support of the European Open Science Cloud, among others, structured FAIRification measures have already been taken for German health datasets.

3.
NPJ Digit Med ; 7(1): 76, 2024 Mar 20.
Artigo em Inglês | MEDLINE | ID: mdl-38509224

RESUMO

Clinical research relies on high-quality patient data, however, obtaining big data sets is costly and access to existing data is often hindered by privacy and regulatory concerns. Synthetic data generation holds the promise of effectively bypassing these boundaries allowing for simplified data accessibility and the prospect of synthetic control cohorts. We employed two different methodologies of generative artificial intelligence - CTAB-GAN+ and normalizing flows (NFlow) - to synthesize patient data derived from 1606 patients with acute myeloid leukemia, a heterogeneous hematological malignancy, that were treated within four multicenter clinical trials. Both generative models accurately captured distributions of demographic, laboratory, molecular and cytogenetic variables, as well as patient outcomes yielding high performance scores regarding fidelity and usability of both synthetic cohorts (n = 1606 each). Survival analysis demonstrated close resemblance of survival curves between original and synthetic cohorts. Inter-variable relationships were preserved in univariable outcome analysis enabling explorative analysis in our synthetic data. Additionally, training sample privacy is safeguarded mitigating possible patient re-identification, which we quantified using Hamming distances. We provide not only a proof-of-concept for synthetic data generation in multimodal clinical data for rare diseases, but also full public access to synthetic data sets to foster further research.

4.
BMC Med Inform Decis Mak ; 24(1): 58, 2024 Feb 26.
Artigo em Inglês | MEDLINE | ID: mdl-38408983

RESUMO

BACKGROUND: To gain insight into the real-life care of patients in the healthcare system, data from hospital information systems and insurance systems are required. Consequently, linking clinical data with claims data is necessary. To ensure their syntactic and semantic interoperability, the Observational Medical Outcomes Partnership (OMOP) Common Data Model (CDM) from the Observational Health Data Sciences and Informatics (OHDSI) community was chosen. However, there is no detailed guide that would allow researchers to follow a generic process for data harmonization, i.e. the transformation of local source data into the standardized OMOP CDM format. Thus, the aim of this paper is to conceptualize a generic data harmonization process for OMOP CDM. METHODS: For this purpose, we conducted a literature review focusing on publications that address the harmonization of clinical or claims data in OMOP CDM. Subsequently, the process steps used and their chronological order as well as applied OHDSI tools were extracted for each included publication. The results were then compared to derive a generic sequence of the process steps. RESULTS: From 23 publications included, a generic data harmonization process for OMOP CDM was conceptualized, consisting of nine process steps: dataset specification, data profiling, vocabulary identification, coverage analysis of vocabularies, semantic mapping, structural mapping, extract-transform-load-process, qualitative and quantitative data quality analysis. Furthermore, we identified seven OHDSI tools which supported five of the process steps. CONCLUSIONS: The generic data harmonization process can be used as a step-by-step guide to assist other researchers in harmonizing source data in OMOP CDM.


Assuntos
Informática Médica , Vocabulário , Humanos , Bases de Dados Factuais , Ciência de Dados , Semântica , Registros Eletrônicos de Saúde
5.
JMIR Med Inform ; 12: e52967, 2024 Feb 14.
Artigo em Inglês | MEDLINE | ID: mdl-38354027

RESUMO

BACKGROUND: Multisite clinical studies are increasingly using real-world data to gain real-world evidence. However, due to the heterogeneity of source data, it is difficult to analyze such data in a unified way across clinics. Therefore, the implementation of Extract-Transform-Load (ETL) or Extract-Load-Transform (ELT) processes for harmonizing local health data is necessary, in order to guarantee the data quality for research. However, the development of such processes is time-consuming and unsustainable. A promising way to ease this is the generalization of ETL/ELT processes. OBJECTIVE: In this work, we investigate existing possibilities for the development of generic ETL/ELT processes. Particularly, we focus on approaches with low development complexity by using descriptive metadata and structural metadata. METHODS: We conducted a literature review following the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines. We used 4 publication databases (ie, PubMed, IEEE Explore, Web of Science, and Biomed Center) to search for relevant publications from 2012 to 2022. The PRISMA flow was then visualized using an R-based tool (Evidence Synthesis Hackathon). All relevant contents of the publications were extracted into a spreadsheet for further analysis and visualization. RESULTS: Regarding the PRISMA guidelines, we included 33 publications in this literature review. All included publications were categorized into 7 different focus groups (ie, medicine, data warehouse, big data, industry, geoinformatics, archaeology, and military). Based on the extracted data, ontology-based and rule-based approaches were the 2 most used approaches in different thematic categories. Different approaches and tools were chosen to achieve different purposes within the use cases. CONCLUSIONS: Our literature review shows that using metadata-driven (MDD) approaches to develop an ETL/ELT process can serve different purposes in different thematic categories. The results show that it is promising to implement an ETL/ELT process by applying MDD approach to automate the data transformation from Fast Healthcare Interoperability Resources to Observational Medical Outcomes Partnership Common Data Model. However, the determining of an appropriate MDD approach and tool to implement such an ETL/ELT process remains a challenge. This is due to the lack of comprehensive insight into the characterizations of the MDD approaches presented in this study. Therefore, our next step is to evaluate the MDD approaches presented in this study and to determine the most appropriate MDD approaches and the way to integrate them into the ETL/ELT process. This could verify the ability of using MDD approaches to generalize the ETL process for harmonizing medical data.

6.
Stud Health Technol Inform ; 310: 1271-1275, 2024 Jan 25.
Artigo em Inglês | MEDLINE | ID: mdl-38270019

RESUMO

To understand and handle the COVID-19 pandemic, digital tools and infrastructures were built in very short timeframes, resulting in stand-alone and non-interoperable solutions. To shape an interoperable, sustainable, and extensible ecosystem to advance biomedical research and healthcare during the pandemic and beyond, a short-term project called "Collaborative Data Exchange and Usage" (CODEX+) was initiated to integrate and connect multiple COVID-19 projects into a common organizational and technical framework. In this paper, we present the conceptual design, provide an overview of the results, and discuss the impact of such a project for the trade-off between innovation and sustainable infrastructures.


Assuntos
Pesquisa Biomédica , COVID-19 , Humanos , Centros Médicos Acadêmicos , COVID-19/epidemiologia , Instalações de Saúde , Pandemias
7.
Sci Rep ; 14(1): 2287, 2024 01 27.
Artigo em Inglês | MEDLINE | ID: mdl-38280887

RESUMO

The emergence of collaborations, which standardize and combine multiple clinical databases across different regions, provide a wealthy source of data, which is fundamental for clinical prediction models, such as patient-level predictions. With the aid of such large data pools, researchers are able to develop clinical prediction models for improved disease classification, risk assessment, and beyond. To fully utilize this potential, Machine Learning (ML) methods are commonly required to process these large amounts of data on disease-specific patient cohorts. As a consequence, the Observational Health Data Sciences and Informatics (OHDSI) collaborative develops a framework to facilitate the application of ML models for these standardized patient datasets by using the Observational Medical Outcomes Partnership (OMOP) common data model (CDM). In this study, we compare the feasibility of current web-based OHDSI approaches, namely ATLAS and "Patient-level Prediction" (PLP), against a native solution (R based) to conduct such ML-based patient-level prediction analyses in OMOP. This will enable potential users to select the most suitable approach for their investigation. Each of the applied ML solutions was individually utilized to solve the same patient-level prediction task. Both approaches went through an exemplary benchmarking analysis to assess the weaknesses and strengths of the PLP R-Package. In this work, the performance of this package was subsequently compared versus the commonly used native R-package called Machine Learning in R 3 (mlr3), and its sub-packages. The approaches were evaluated on performance, execution time, and ease of model implementation. The results show that the PLP package has shorter execution times, which indicates great scalability, as well as intuitive code implementation, and numerous possibilities for visualization. However, limitations in comparison to native packages were depicted in the implementation of specific ML classifiers (e.g., Lasso), which may result in a decreased performance for real-world prediction problems. The findings here contribute to the overall effort of developing ML-based prediction models on a clinical scale and provide a snapshot for future studies that explicitly aim to develop patient-level prediction models in OMOP CDM.


Assuntos
Aprendizado de Máquina , Informática Médica , Humanos , Bases de Dados Factuais , Registros Eletrônicos de Saúde
8.
PLoS One ; 19(1): e0297039, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38295046

RESUMO

BACKGROUND: The COVID-19 pandemic revealed a need for better collaboration among research, care, and management in Germany as well as globally. Initially, there was a high demand for broad data collection across Germany, but as the pandemic evolved, localized data became increasingly necessary. Customized dashboards and tools were rapidly developed to provide timely and accurate information. In Saxony, the DISPENSE project was created to predict short-term hospital bed capacity demands, and while it was successful, continuous adjustments and the initial monolithic system architecture of the application made it difficult to customize and scale. METHODS: To analyze the current state of the DISPENSE tool, we conducted an in-depth analysis of the data processing steps and identified data flows underlying users' metrics and dashboards. We also conducted a workshop to understand the different views and constraints of specific user groups, and brought together and clustered the information according to content-related service areas to determine functionality-related service groups. Based on this analysis, we developed a concept for the system architecture, modularized the main services by assigning specialized applications and integrated them into the existing system, allowing for self-service reporting and evaluation of the expert groups' needs. RESULTS: We analyzed the applications' dataflow and identified specific user groups. The functionalities of the monolithic application were divided into specific service groups for data processing, data storage, predictions, content visualization, and user management. After composition and implementation, we evaluated the new system architecture against the initial requirements by enabling self-service reporting to the users. DISCUSSION: By modularizing the monolithic application and creating a more flexible system, the challenges of rapidly changing requirements, growing need for information, and high administrative efforts were addressed. CONCLUSION: We demonstrated an improved adaptation towards the needs of various user groups, increased efficiency, and reduced burden on administrators, while also enabling self-service functionalities and specialization of single applications on individual service groups.


Assuntos
Armazenamento e Recuperação da Informação , Pandemias , Humanos , Coleta de Dados , Alemanha
9.
JMIR Med Inform ; 11: e47959, 2023 Nov 07.
Artigo em Inglês | MEDLINE | ID: mdl-37942786

RESUMO

Background: National classifications and terminologies already routinely used for documentation within patient care settings enable the unambiguous representation of clinical information. However, the diversity of different vocabularies across health care institutions and countries is a barrier to achieving semantic interoperability and exchanging data across sites. The Observational Medical Outcomes Partnership (OMOP) Common Data Model (CDM) enables the standardization of structure and medical terminology. It allows the mapping of national vocabularies into so-called standard concepts, representing normative expressions for international analyses and research. Within our project "Hybrid Quality Indicators Using Machine Learning Methods" (Hybrid-QI), we aim to harmonize source codes used in German claims data vocabularies that are currently unavailable in the OMOP CDM. Objective: This study aims to increase the coverage of German vocabularies in the OMOP CDM. We aim to completely transform the source codes used in German claims data into the OMOP CDM without data loss and make German claims data usable for OMOP CDM-based research. Methods: To prepare the missing German vocabularies for the OMOP CDM, we defined a vocabulary preparation approach consisting of the identification of all codes of the corresponding vocabularies, their assembly into machine-readable tables, and the translation of German designations into English. Furthermore, we used 2 proposed approaches for OMOP-compliant vocabulary preparation: the mapping to standard concepts using the Observational Health Data Sciences and Informatics (OHDSI) tool Usagi and the preparation of new 2-billion concepts (ie, concept_id >2 billion). Finally, we evaluated the prepared vocabularies regarding completeness and correctness using synthetic German claims data and calculated the coverage of German claims data vocabularies in the OMOP CDM. Results: Our vocabulary preparation approach was able to map 3 missing German vocabularies to standard concepts and prepare 8 vocabularies as new 2-billion concepts. The completeness evaluation showed that the prepared vocabularies cover 44.3% (3288/7417) of the source codes contained in German claims data. The correctness evaluation revealed that the specified validity periods in the OMOP CDM are compliant for the majority (705,531/706,032, 99.9%) of source codes and associated dates in German claims data. The calculation of the vocabulary coverage showed a noticeable decrease of missing vocabularies from 55% (11/20) to 10% (2/20) due to our preparation approach. Conclusions: By preparing 10 vocabularies, we showed that our approach is applicable to any type of vocabulary used in a source data set. The prepared vocabularies are currently limited to German vocabularies, which can only be used in national OMOP CDM research projects, because the mapping of new 2-billion concepts to standard concepts is missing. To participate in international OHDSI network studies with German claims data, future work is required to map the prepared 2-billion concepts to standard concepts.

10.
JMIR Med Inform ; 11: e45116, 2023 Aug 03.
Artigo em Inglês | MEDLINE | ID: mdl-37535410

RESUMO

BACKGROUND: Common data models (CDMs) are essential tools for data harmonization, which can lead to significant improvements in the health domain. CDMs unite data from disparate sources and ease collaborations across institutions, resulting in the generation of large standardized data repositories across different entities. An overview of existing CDMs and methods used to develop these data sets may assist in the development process of future models for the health domain, such as for decision support systems. OBJECTIVE: This scoping review investigates methods used in the development of CDMs for health data. We aim to provide a broad overview of approaches and guidelines that are used in the development of CDMs (ie, common data elements or common data sets) for different health domains on an international level. METHODS: This scoping review followed the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) checklist. We conducted the literature search in prominent databases, namely, PubMed, Web of Science, Science Direct, and Scopus, starting from January 2000 until March 2022. We identified and screened 1309 articles. The included articles were evaluated based on the type of adopted method, which was used in the conception, users' needs collection, implementation, and evaluation phases of CDMs, and whether stakeholders (such as medical experts, patients' representatives, and IT staff) were involved during the process. Moreover, the models were grouped into iterative or linear types based on the imperativeness of the stages during development. RESULTS: We finally identified 59 articles that fit our eligibility criteria. Of these articles, 45 specifically focused on common medical conditions, 10 focused on rare medical conditions, and the remaining 4 focused on both conditions. The development process usually involved stakeholders but in different ways (eg, working group meetings, Delphi approaches, interviews, and questionnaires). Twenty-two models followed an iterative process. CONCLUSIONS: The included articles showed the diversity of methods used to develop a CDM in different domains of health. We highlight the need for more specialized CDM development methods in the health domain and propose a suggestive development process that might ease the development of CDMs in the health domain in the future.

11.
JMIR Med Inform ; 11: e47310, 2023 Aug 21.
Artigo em Inglês | MEDLINE | ID: mdl-37621207

RESUMO

Background: In the Medical Informatics in Research and Care in University Medicine (MIRACUM) consortium, an IT-based clinical trial recruitment support system was developed based on the Observational Medical Outcomes Partnership (OMOP) Common Data Model (CDM). Currently, OMOP CDM is populated with German Fast Healthcare Interoperability Resources (FHIR) using an Extract-Transform-Load (ETL) process, which was designed as a bulk load. However, the computational effort that comes with an everyday full load is not efficient for daily recruitment. Objective: The aim of this study is to extend our existing ETL process with the option of incremental loading to efficiently support daily updated data. Methods: Based on our existing bulk ETL process, we performed an analysis to determine the requirements of incremental loading. Furthermore, a literature review was conducted to identify adaptable approaches. Based on this, we implemented three methods to integrate incremental loading into our ETL process. Lastly, a test suite was defined to evaluate the incremental loading for data correctness and performance compared to bulk loading. Results: The resulting ETL process supports bulk and incremental loading. Performance tests show that the incremental load took 87.5% less execution time than the bulk load (2.12 min compared to 17.07 min) related to changes of 1 day, while no data differences occurred in OMOP CDM. Conclusions: Since incremental loading is more efficient than a daily bulk load and both loading options result in the same amount of data, we recommend using bulk load for an initial load and switching to incremental load for daily updates. The resulting incremental ETL logic can be applied internationally since it is not restricted to German FHIR profiles.

12.
J Med Internet Res ; 25: e45948, 2023 07 24.
Artigo em Inglês | MEDLINE | ID: mdl-37486754

RESUMO

The vast and heterogeneous data being constantly generated in clinics can provide great wealth for patients and research alike. The quickly evolving field of medical informatics research has contributed numerous concepts, algorithms, and standards to facilitate this development. However, these difficult relationships, complex terminologies, and multiple implementations can present obstacles for people who want to get active in the field. With a particular focus on medical informatics research conducted in Germany, we present in our Viewpoint a set of 10 important topics to improve the overall interdisciplinary communication between different stakeholders (eg, physicians, computational experts, experimentalists, students, patient representatives). This may lower the barriers to entry and offer a starting point for collaborations at different levels. The suggested topics are briefly introduced, then general best practice guidance is given, and further resources for in-depth reading or hands-on tutorials are recommended. In addition, the topics are set to cover current aspects and open research gaps of the medical informatics domain, including data regulations and concepts; data harmonization and processing; and data evaluation, visualization, and dissemination. In addition, we give an example on how these topics can be integrated in a medical informatics curriculum for higher education. By recognizing these topics, readers will be able to (1) set clinical and research data into the context of medical informatics, understanding what is possible to achieve with data or how data should be handled in terms of data privacy and storage; (2) distinguish current interoperability standards and obtain first insights into the processes leading to effective data transfer and analysis; and (3) value the use of newly developed technical approaches to utilize the full potential of clinical data.


Assuntos
Informática Médica , Humanos , Currículo , Algoritmos , Alemanha
13.
Stud Health Technol Inform ; 305: 139-140, 2023 Jun 29.
Artigo em Inglês | MEDLINE | ID: mdl-37386977

RESUMO

Current challenges of rare diseases need to involve patients, physicians, and the research community to generate new insights on comprehensive patient cohorts. Interestingly, the integration of patient context has been insufficiently considered, but might tremendously improve the accuracy of predictive models for individual patients. Here, we conceptualized an extension of the European Platform for Rare Disease Registration data model with contextual factors. This extended model can serve as an enhanced baseline and is well-suited for analyses using artificial intelligence models for improved predictions. The study is an initial result that will develop context-sensitive common data models for genetic rare diseases.


Assuntos
Inteligência Artificial , Médicos , Humanos , Doenças Raras/genética
14.
Stud Health Technol Inform ; 302: 711-715, 2023 May 18.
Artigo em Inglês | MEDLINE | ID: mdl-37203475

RESUMO

INTRODUCTION: Real-world data (RWD) is gaining importance in research. For instance, the European Medicines Agency (EMA) is currently in the process of establishing a cross-national research network that utilizes RWD for research. However, data harmonization across countries must be carefully considered to avoid misclassification and bias. OBJECTIVES: This paper aims to investigate the extent to which a correct assignment of RxNorm ingredients is possible for medication orders that include only ATC codes. METHODS: In this study, we analyzed 1,506,059 medication orders from the University Hospital Dresden (UKD) and merged them with the ATC vocabulary in the Observational Medical Outcomes Partnership (OMOP) including relevant relationship mappings to RxNorm. RESULTS: We identified 70.25% of all medication orders were single ingredients with direct mapping to RxNorm. However, we also identified a significant complexity in mappings for the other medication orders that was visualized in an interactive scatterplot. DISCUSSION: The majority of medication orders under observation (70.25%) are single ingredients and can be standardized to RxNorm, combination drugs pose a challenge due to the different approaches of ingredient assignments in ATC and RxNorm. The provided visualization can help research teams gain a better understanding of problematic data and further investigate identified issues.


Assuntos
RxNorm , Vocabulário Controlado , Humanos , Registros , Vocabulário , Hospitais Universitários
15.
Stud Health Technol Inform ; 302: 751-752, 2023 May 18.
Artigo em Inglês | MEDLINE | ID: mdl-37203486

RESUMO

OMOP common data model (CDM) is designed for analyzing large clinical data and building cohorts for medical research, which requires Extract-Transform-Load processes (ETL) of local heterogeneous medical data. We present a concept for developing and evaluating a modularized metadata-driven ETL process, which can transform data into OMOP CDM regardless of 1) the source data format, 2) its versions and 3) context of use.


Assuntos
Pesquisa Biomédica , Metadados , Registros Eletrônicos de Saúde , Bases de Dados Factuais
16.
Stud Health Technol Inform ; 302: 753-754, 2023 May 18.
Artigo em Inglês | MEDLINE | ID: mdl-37203487

RESUMO

The availability of clinical data for researchers is crucial for an improvement of healthcare and research. For this purpose, the integration, harmonization and standardization of healthcare-data from various sources in a clinical data warehouse (CDWH) is highly relevant. Our evaluation taking into account the general conditions and requirements of the project, led us to choose the Data Vault approach for the development of a clinical data warehouse at the University Hospital Dresden (UHD).


Assuntos
Data Warehousing , Atenção à Saúde , Humanos , Instalações de Saúde , Padrões de Referência
17.
Stud Health Technol Inform ; 302: 3-7, 2023 May 18.
Artigo em Inglês | MEDLINE | ID: mdl-37203598

RESUMO

Research on real-world data is becoming increasingly important. The current restriction to clinical data in Germany limits the view of the patient. To gain comprehensive insights, claims data can be added to the existing knowledge. However, standardized transfer of German claims data into OMOP CDM is currently not possible. In this paper, we conducted an evaluation regarding the coverage of source vocabularies and data elements of German claims data in OMOP CDM. We point out the need to extend vocabularies and mappings to support research on German claims data.


Assuntos
Registros Eletrônicos de Saúde , Vocabulário , Humanos , Alemanha , Bases de Dados Factuais
18.
JMIR Med Inform ; 11: e40312, 2023 Jan 25.
Artigo em Inglês | MEDLINE | ID: mdl-36696159

RESUMO

BACKGROUND: Digitization offers a multitude of opportunities to gain insights into current diagnostics and therapies from retrospective data. In this context, real-world data and their accessibility are of increasing importance to support unbiased and reliable research on big data. However, routinely collected data are not readily usable for research owing to the unstructured nature of health care systems and a lack of interoperability between these systems. This challenge is evident in drug data. OBJECTIVE: This study aimed to present an approach that identifies and increases the structuredness of drug data while ensuring standardization according to Anatomical Therapeutic Chemical (ATC) classification. METHODS: Our approach was based on available drug prescriptions and a drug catalog and consisted of 4 steps. First, we performed an initial analysis of the structuredness of local drug data to define a point of comparison for the effectiveness of the overall approach. Second, we applied 3 algorithms to unstructured data that translated text into ATC codes based on string comparisons in terms of ingredients and product names and performed similarity comparisons based on Levenshtein distance. Third, we validated the results of the 3 algorithms with expert knowledge based on the 1000 most frequently used prescription texts. Fourth, we performed a final validation to determine the increased degree of structuredness. RESULTS: Initially, 47.73% (n=843,980) of 1,768,153 drug prescriptions were classified as structured. With the application of the 3 algorithms, we were able to increase the degree of structuredness to 85.18% (n=1,506,059) based on the 1000 most frequent medication prescriptions. In this regard, the combination of algorithms 1, 2, and 3 resulted in a correctness level of 100% (with 57,264 ATC codes identified), algorithms 1 and 3 resulted in 99.6% (with 152,404 codes identified), and algorithms 1 and 2 resulted in 95.9% (with 39,472 codes identified). CONCLUSIONS: As shown in the first analysis steps of our approach, the availability of a product catalog to select during the documentation process is not sufficient to generate structured data. Our 4-step approach reduces the problems and reliably increases the structuredness automatically. Similarity matching shows promising results, particularly for entries with no connection to a product catalog. However, further enhancement of the correctness of such a similarity matching algorithm needs to be investigated in future work.

19.
Int J Med Inform ; 169: 104925, 2023 01.
Artigo em Inglês | MEDLINE | ID: mdl-36395615

RESUMO

BACKGROUND: International studies are increasingly needed in order to gain more unbiased evidence from real-world data. To achieve this goal across the European Union, the EMA set up the DARWIN EU project based on OMOP CDM established by the OHDSI community. The harmonization of heterogeneous local health data in OMOP CDM is an essential step to participate in such networks. Using the widespread communication standard HL7 FHIR can reduce the complexity of the transformation process to OMOP CDM. Enabling German university hospitals to participate in such networks requires an Extract, Transform and Load (ETL)-process that satisfies the following criteria: 1) transforming German patient data from FHIR to OMOP CDM, 2) processing huge amount of data at once and 3) flexibility to cope with changes in FHIR profiles. METHOD: A mapping of German patient data from FHIR to OMOP CDM was accomplished, validated by an interdisciplinary team and checked through the OHDSI Data Quality Dashboard (DQD). To satisfy criteria 2-3, we decided to use SpringBatch-Framework according to its chunk-oriented design and reusable functions for processing large amounts of data. RESULTS: We have successfully developed an ETL-process that fulfills the defined criteria of transforming German patient data from FHIR into OMOP CDM. To measure the validity of the mapping conformance and performance of the ETL-process, it was tested with 392,022 FHIR resources. The ETL execution lasted approximately-one minute and the DQD result shows 99% conformance in OMOP CDM. CONCLUSIONS: Our ETL-process has been successfully tested and integrated at 10 German university hospitals. The data harmonization utilizing international recognized standards like FHIR and OMOP fosters their ability to participate in international observational studies. Additionally, the ETL process can help to prepare more German hospitals with their data harmonization journey based on existing standards.


Assuntos
Mineração de Dados , Cooperação Internacional , Humanos , União Europeia
20.
Digit Health ; 8: 20552076221134437, 2022.
Artigo em Inglês | MEDLINE | ID: mdl-36325436

RESUMO

Objective: Limited capacities and ineffective care pathways result in long waiting times for patients and sporadic treatment controls in sleep medicine. As one objective of the 'Telesleep Medicine' project, a portal should be developed, which supports sleep specialists in an efficient and resource-saving patient management. On account of the limited project timeframe, the 'classical' user-centred design and evaluation methods could not be comprehensively implemented. Therefore, a pragmatic methodical framework was developed. Methods: For the iterative development of the portal, a combination of low-cost and quick-to-implement methods was used. In chronological order, these were: context interviews, personas, the development of an as-is model, a web search of design standards and good design aspects of similar systems, the development of a to-be model, the creation of an overarching mind map, and the iterative creation of mockups with simplified usability walkthroughs. Results: The feasibility of the pragmatic methodological framework for the development of a prototype for the portal was demonstrated. The used method combination resulted in a prototype based on the needs and requirements of the sleep specialists, taking into account their specific workflow and the technical implementation conditions. Conclusions: The presented pragmatic methodological framework can be a valuable resource for developers of comparable projects. The combination of methods worked well together regarding the limited timeframe and resources for concept development. For the future, we plan to implement and test the portal in the clinical field and thus enrich our framework with additional methods.

SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA