Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 21
Filtrar
Mais filtros

Bases de dados
País/Região como assunto
Tipo de documento
Intervalo de ano de publicação
1.
Neurorehabil Neural Repair ; 37(9): 591-602, 2023 09.
Artigo em Inglês | MEDLINE | ID: mdl-37592867

RESUMO

BACKGROUND: The incidence of stroke and stroke-related hemiparesis has been steadily increasing and is projected to become a serious social, financial, and physical burden on the aging population. Limited access to outpatient rehabilitation for these stroke survivors further deepens the healthcare issue and estranges the stroke patient demographic in rural areas. However, new advances in motion detection deep learning enable the use of handheld smartphone cameras for body tracking, offering unparalleled levels of accessibility. METHODS: In this study we want to develop an automated method for evaluation of a shortened variant of the Fugl-Meyer assessment, the standard stroke rehabilitation scale describing upper extremity motor function. We pair this technology with a series of machine learning models, including different neural network structures and an eXtreme Gradient Boosting model, to score 16 of 33 (49%) Fugl-Meyer item activities. RESULTS: In this observational study, 45 acute stroke patients completed at least 1 recorded Fugl-Meyer assessment for the training of the auto-scorers, which yielded average accuracies ranging from 78.1% to 82.7% item-wise. CONCLUSION: In this study, an automated method was developed for the evaluation of a shortened variant of the Fugl-Meyer assessment, the standard stroke rehabilitation scale describing upper extremity motor function. This novel method is demonstrated with potential to conduct telehealth rehabilitation evaluations and assessments with accuracy and availability.


Assuntos
Reabilitação do Acidente Vascular Cerebral , Acidente Vascular Cerebral , Humanos , Idoso , Captura de Movimento , Biônica , Recuperação de Função Fisiológica , Avaliação da Deficiência , Acidente Vascular Cerebral/complicações , Acidente Vascular Cerebral/diagnóstico , Reabilitação do Acidente Vascular Cerebral/métodos , Extremidade Superior
2.
J Am Med Inform Assoc ; 30(9): 1465-1473, 2023 08 18.
Artigo em Inglês | MEDLINE | ID: mdl-37301740

RESUMO

OBJECTIVE: Social determinants of health (SDoH) play critical roles in health outcomes and well-being. Understanding the interplay of SDoH and health outcomes is critical to reducing healthcare inequalities and transforming a "sick care" system into a "health-promoting" system. To address the SDOH terminology gap and better embed relevant elements in advanced biomedical informatics, we propose an SDoH ontology (SDoHO), which represents fundamental SDoH factors and their relationships in a standardized and measurable way. MATERIAL AND METHODS: Drawing on the content of existing ontologies relevant to certain aspects of SDoH, we used a top-down approach to formally model classes, relationships, and constraints based on multiple SDoH-related resources. Expert review and coverage evaluation, using a bottom-up approach employing clinical notes data and a national survey, were performed. RESULTS: We constructed the SDoHO with 708 classes, 106 object properties, and 20 data properties, with 1,561 logical axioms and 976 declaration axioms in the current version. Three experts achieved 0.967 agreement in the semantic evaluation of the ontology. A comparison between the coverage of the ontology and SDOH concepts in 2 sets of clinical notes and a national survey instrument also showed satisfactory results. DISCUSSION: SDoHO could potentially play an essential role in providing a foundation for a comprehensive understanding of the associations between SDoH and health outcomes and paving the way for health equity across populations. CONCLUSION: SDoHO has well-designed hierarchies, practical objective properties, and versatile functionalities, and the comprehensive semantic and coverage evaluation achieved promising performance compared to the existing ontologies relevant to SDoH.


Assuntos
Equidade em Saúde , Determinantes Sociais da Saúde , Humanos , Semântica , Disparidades em Assistência à Saúde
3.
J Biomed Inform ; 139: 104269, 2023 03.
Artigo em Inglês | MEDLINE | ID: mdl-36621750

RESUMO

Electronic health records (EHR) are collected as a routine part of healthcare delivery, and have great potential to be utilized to improve patient health outcomes. They contain multiple years of health information to be leveraged for risk prediction, disease detection, and treatment evaluation. However, they do not have a consistent, standardized format across institutions, particularly in the United States, and can present significant analytical challenges- they contain multi-scale data from heterogeneous domains and include both structured and unstructured data. Data for individual patients are collected at irregular time intervals and with varying frequencies. In addition to the analytical challenges, EHR can reflect inequity- patients belonging to different groups will have differing amounts of data in their health records. Many of these issues can contribute to biased data collection. The consequence is that the data for under-served groups may be less informative partly due to more fragmented care, which can be viewed as a type of missing data problem. For EHR data in this complex form, there is currently no framework for introducing realistic missing values. There has also been little to no work in assessing the impact of missing data in EHR. In this work, we first introduce a terminology to define three levels of EHR data and then propose a novel framework for simulating realistic missing data scenarios in EHR to adequately assess their impact on predictive modeling. We incorporate the use of a medical knowledge graph to capture dependencies between medical events to create a more realistic missing data framework. In an intensive care unit setting, we found that missing data have greater negative impact on the performance of disease prediction models in groups that tend to have less access to healthcare, or seek less healthcare. We also found that the impact of missing data on disease prediction models is stronger when using the knowledge graph framework to introduce realistic missing values as opposed to random event removal.


Assuntos
Atenção à Saúde , Registros Eletrônicos de Saúde , Humanos , Estados Unidos , Unidades de Terapia Intensiva
4.
AMIA Annu Symp Proc ; 2023: 814-823, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-38222389

RESUMO

In the era of big data, there is an increasing need for healthcare providers, communities, and researchers to share data and collaborate to improve health outcomes, generate valuable insights, and advance research. The Health Insurance Portability and Accountability Act of 1996 (HIPAA) is a federal law designed to protect sensitive health information by defining regulations for protected health information (PHI). However, it does not provide efficient tools for detecting or removing PHI before data sharing. One of the challenges in this area of research is the heterogeneous nature of PHI fields in data across different parties. This variability makes rule-based sensitive variable identification systems that work on one database fail on another. To address this issue, our paper explores the use of machine learning algorithms to identify sensitive variables in structured data, thus facilitating the de-identification process. We made a key observation that the distributions of metadata of PHI fields and non-PHI fields are very different. Based on this novel finding, we engineered over 30 features from the metadata of the original features and used machine learning to build classification models to automatically identify PHI fields in structured Electronic Health Record (EHR) data. We trained the model on a variety of large EHR databases from different data sources and found that our algorithm achieves 99% accuracy when detecting PHI-related fields for unseen datasets. The implications of our study are significant and can benefit industries that handle sensitive data.


Assuntos
Confidencialidade , Sistemas Computadorizados de Registros Médicos , Estados Unidos , Humanos , Health Insurance Portability and Accountability Act , Algoritmos , Aprendizado de Máquina , Registros Eletrônicos de Saúde
5.
BMC Bioinformatics ; 23(1): 409, 2022 Oct 01.
Artigo em Inglês | MEDLINE | ID: mdl-36182914

RESUMO

BACKGROUND: Sequencing of thousands of samples provides genetic variants with allele frequencies spanning a very large spectrum and gives invaluable insight into genetic determinants of diseases. Protecting the genetic privacy of participants is challenging as only a few rare variants can easily re-identify an individual among millions. In certain cases, there are policy barriers against sharing genetic data from indigenous populations and stigmatizing conditions. RESULTS: We present SVAT, a method for secure outsourcing of variant annotation and aggregation, which are two basic steps in variant interpretation and detection of causal variants. SVAT uses homomorphic encryption to encrypt the data at the client-side. The data always stays encrypted while it is stored, in-transit, and most importantly while it is analyzed. SVAT makes use of a vectorized data representation to convert annotation and aggregation into efficient vectorized operations in a single framework. Also, SVAT utilizes a secure re-encryption approach so that multiple disparate genotype datasets can be combined for federated aggregation and secure computation of allele frequencies on the aggregated dataset. CONCLUSIONS: Overall, SVAT provides a secure, flexible, and practical framework for privacy-aware outsourcing of annotation, filtering, and aggregation of genetic variants. SVAT is publicly available for download from https://github.com/harmancilab/SVAT .


Assuntos
Computação em Nuvem , Serviços Terceirizados , Segurança Computacional , Frequência do Gene , Genótipo , Humanos
6.
BMC Bioinformatics ; 23(1): 356, 2022 Aug 29.
Artigo em Inglês | MEDLINE | ID: mdl-36038834

RESUMO

BACKGROUND: The decreasing cost of DNA sequencing has led to a great increase in our knowledge about genetic variation. While population-scale projects bring important insight into genotype-phenotype relationships, the cost of performing whole-genome sequencing on large samples is still prohibitive. In-silico genotype imputation coupled with genotyping-by-arrays is a cost-effective and accurate alternative for genotyping of common and uncommon variants. Imputation methods compare the genotypes of the typed variants with the large population-specific reference panels and estimate the genotypes of untyped variants by making use of the linkage disequilibrium patterns. Most accurate imputation methods are based on the Li-Stephens hidden Markov model, HMM, that treats the sequence of each chromosome as a mosaic of the haplotypes from the reference panel. RESULTS: Here we assess the accuracy of vicinity-based HMMs, where each untyped variant is imputed using the typed variants in a small window around itself (as small as 1 centimorgan). Locality-based imputation is used recently by machine learning-based genotype imputation approaches. We assess how the parameters of the vicinity-based HMMs impact the imputation accuracy in a comprehensive set of benchmarks and show that vicinity-based HMMs can accurately impute common and uncommon variants. CONCLUSIONS: Our results indicate that locality-based imputation models can be effectively used for genotype imputation. The parameter settings that we identified can be used in future methods and vicinity-based HMMs can be used for re-structuring and parallelizing new imputation methods. The source code for the vicinity-based HMM implementations is publicly available at https://github.com/harmancilab/LoHaMMer .


Assuntos
Polimorfismo de Nucleotídeo Único , Software , Estudo de Associação Genômica Ampla/métodos , Genótipo , Haplótipos , Desequilíbrio de Ligação , Análise de Sequência de DNA/métodos
7.
Cell Syst ; 12(11): 1108-1120.e4, 2021 11 17.
Artigo em Inglês | MEDLINE | ID: mdl-34464590

RESUMO

Genotype imputation is a fundamental step in genomic data analysis, where missing variant genotypes are predicted using the existing genotypes of nearby "tag" variants. Although researchers can outsource genotype imputation, privacy concerns may prohibit genetic data sharing with an untrusted imputation service. Here, we developed secure genotype imputation using efficient homomorphic encryption (HE) techniques. In HE-based methods, the genotype data are secure while it is in transit, at rest, and in analysis. It can only be decrypted by the owner. We compared secure imputation with three state-of-the-art non-secure methods and found that HE-based methods provide genetic data security with comparable accuracy for common variants. HE-based methods have time and memory requirements that are comparable or lower than those for the non-secure methods. Our results provide evidence that HE-based methods can practically perform resource-intensive computations for high-throughput genetic data analysis. The source code is freely available for download at https://github.com/K-miran/secure-imputation.


Assuntos
Serviços Terceirizados , Segurança Computacional , Estudo de Associação Genômica Ampla , Genótipo , Privacidade
8.
Sci Rep ; 11(1): 9313, 2021 04 29.
Artigo em Inglês | MEDLINE | ID: mdl-33927277

RESUMO

Our objective is to derive a sequential decision-making rule on the combination of medications to minimize motor symptoms using reinforcement learning (RL). Using an observational longitudinal cohort of Parkinson's disease patients, the Parkinson's Progression Markers Initiative database, we derived clinically relevant disease states and an optimal combination of medications for each of them by using policy iteration of the Markov decision process (MDP). We focused on 8 combinations of medications, i.e., Levodopa, a dopamine agonist, and other PD medications, as possible actions and motor symptom severity, based on the Unified Parkinson Disease Rating Scale (UPDRS) section III, as reward/penalty of decision. We analyzed a total of 5077 visits from 431 PD patients with 55.5 months follow-up. We excluded patients without UPDRS III scores or medication records. We derived a medication regimen that is comparable to a clinician's decision. The RL model achieved a lower level of motor symptom severity scores than what clinicians did, whereas the clinicians' medication rules were more consistent than the RL model. The RL model followed the clinician's medication rules in most cases but also suggested some changes, which leads to the difference in lowering symptoms severity. This is the first study to investigate RL to improve the pharmacological approach of PD patients. Our results contribute to the development of an interactive machine-physician ecosystem that relies on evidence-based medicine and can potentially enhance PD management.


Assuntos
Quimioterapia Combinada , Aprendizado de Máquina , Doença de Parkinson/tratamento farmacológico , Idoso , Estudos de Coortes , Feminino , Humanos , Masculino , Cadeias de Markov , Pessoa de Meia-Idade
9.
iScience ; 24(2): 102106, 2021 Feb 19.
Artigo em Inglês | MEDLINE | ID: mdl-33659874

RESUMO

Sepsis is a leading cause of death among inpatients at hospitals. However, with early detection, death rate can drop substantially. In this study, we present the top-performing algorithm for Sepsis II prediction in the DII National Data Science Challenge using the Cerner Health Facts data involving more than 100,000 adult patients. This large sample size allowed us to dissect the predictability by age-groups, race, genders, and care settings and up to 192 hr of sepsis onset. This large data collection also allowed us to conclude that the last six biometric records on average are informative to the prediction of sepsis. We identified biomarkers that are common across the treatment time and novel biomarkers that are uniquely presented for early prediction. The algorithms showed meaningful signals days ahead of sepsis onset, supporting the potential of reducing death rate by focusing on high-risk populations identified from heterogeneous data integration.

10.
BMC Med Inform Decis Mak ; 20(Suppl 12): 326, 2020 12 24.
Artigo em Inglês | MEDLINE | ID: mdl-33357224

RESUMO

BACKGROUND: Sudden Unexpected Death in Epilepsy (SUDEP) has increased in awareness considerably over the last two decades and is acknowledged as a serious problem in epilepsy. However, the scientific community remains unclear on the reason or possible bio markers that can discern potentially fatal seizures from other non-fatal seizures. The duration of postictal generalized EEG suppression (PGES) is a promising candidate to aid in identifying SUDEP risk. The length of time a patient experiences PGES after a seizure may be used to infer the risk a patient may have of SUDEP later in life. However, the problem becomes identifying the duration, or marking the end, of PGES (Tomson et al. in Lancet Neurol 7(11):1021-1031, 2008; Nashef in Epilepsia 38:6-8, 1997). METHODS: This work addresses the problem of marking the end to PGES in EEG data, extracted from patients during a clinically supervised seizure. This work proposes a sensitivity analysis on EEG window size/delay, feature extraction and classifiers along with associated hyperparameters. The resulting sensitivity analysis includes the Gradient Boosted Decision Trees and Random Forest classifiers trained on 10 extracted features rooted in fundamental EEG behavior using an EEG specific feature extraction process (pyEEG) and 5 different window sizes or delays (Bao et al. in Comput Intell Neurosci 2011:1687-5265, 2011). RESULTS: The machine learning architecture described above scored a maximum AUC score of 76.02% with the Random Forest classifier trained on all extracted features. The highest performing features included SVD Entropy, Petrosan Fractal Dimension and Power Spectral Intensity. CONCLUSION: The methods described are effective in automatically marking the end to PGES. Future work should include integration of these methods into the clinical setting and using the results to be able to predict a patient's SUDEP risk.


Assuntos
Epilepsia , Morte Súbita Inesperada na Epilepsia , Eletroencefalografia , Epilepsia/diagnóstico , Humanos , Medição de Risco , Convulsões/diagnóstico
11.
Sci Total Environ ; 657: 1051-1063, 2019 Mar 20.
Artigo em Inglês | MEDLINE | ID: mdl-30677873

RESUMO

The health and resilience of urban ecosystems are highly dependent on interactions between the natural and built environments. Rapid urbanization, however, brings potential risks to urban ecosystems. Therefore, it is important to justify land-use planning and to identify opportunities for regulating sustainable land use. We developed a framework of land-use planning based on an Environmental Risk Assessment (ERA) and applied it to an environmentally sensitive urban area in Shenzhen, China. The ERA used the analytic hierarchy process method to determine weights for various indicators, and further performed a multifactor-based spatial superposition analysis in the ModelBuilder of a geographic information system to produce a risk map. The selected indicators were topography, hydrology, ecosystem, land use, and traffic. The risk map was first compared with an existing map of ecological control lines to ensure the reliability of the ERA, and then applied to establish the necessity and priority of land-use measures for four different land-use types: nature reserves, green space, urban areas, and spare land. The map indicated that nature reserves and green space make up 84.7% of the area at high risk of degradation, whereas urban areas make up 13.7%. It also showed that a majority of the high-risk urban areas are distributed around water-source reserves and the Pingshan River, and 87.6% of them are residential, industrial, and commercial lands in which the potential risks of non-point source pollution are high. Corrective actions should be considered on an urgent basis in high-risk urban areas, where low-impact development practices are considered effective in reducing non-point source pollution at the source. Validation results affirmed that the proposed ERA approach can reliably provide insights into the distribution of environmental risks in the study area. The proposed framework of ERA-based land-use planning is applicable to sustainable urban development and revitalization of other urban regions.

12.
BMC Med Genomics ; 10(Suppl 2): 48, 2017 07 26.
Artigo em Inglês | MEDLINE | ID: mdl-28786365

RESUMO

BACKGROUND: Advances in DNA sequencing technologies have prompted a wide range of genomic applications to improve healthcare and facilitate biomedical research. However, privacy and security concerns have emerged as a challenge for utilizing cloud computing to handle sensitive genomic data. METHODS: We present one of the first implementations of Software Guard Extension (SGX) based securely outsourced genetic testing framework, which leverages multiple cryptographic protocols and minimal perfect hash scheme to enable efficient and secure data storage and computation outsourcing. RESULTS: We compared the performance of the proposed PRESAGE framework with the state-of-the-art homomorphic encryption scheme, as well as the plaintext implementation. The experimental results demonstrated significant performance over the homomorphic encryption methods and a small computational overhead in comparison to plaintext implementation. CONCLUSIONS: The proposed PRESAGE provides an alternative solution for secure and efficient genomic data outsourcing in an untrusted cloud by using a hybrid framework that combines secure hardware and multiple crypto protocols.


Assuntos
Segurança Computacional , Testes Genéticos , Análise de Sequência de DNA , Software , Computação em Nuvem , Serviços Terceirizados
13.
Ann N Y Acad Sci ; 1387(1): 73-83, 2017 01.
Artigo em Inglês | MEDLINE | ID: mdl-27681358

RESUMO

Accessing and integrating human genomic data with phenotypes are important for biomedical research. Making genomic data accessible for research purposes, however, must be handled carefully to avoid leakage of sensitive individual information to unauthorized parties and improper use of data. In this article, we focus on data sharing within the scope of data accessibility for research. Current common practices to gain biomedical data access are strictly rule based, without a clear and quantitative measurement of the risk of privacy breaches. In addition, several types of studies require privacy-preserving linkage of genotype and phenotype information across different locations (e.g., genotypes stored in a sequencing facility and phenotypes stored in an electronic health record) to accelerate discoveries. The computer science community has developed a spectrum of techniques for data privacy and confidentiality protection, many of which have yet to be tested on real-world problems. In this article, we discuss clinical, technical, and ethical aspects of genome data privacy and confidentiality in the United States, as well as potential solutions for privacy-preserving genotype-phenotype linkage in biomedical research.


Assuntos
Privacidade Genética , Genômica/métodos , Biologia Computacional/ética , Biologia Computacional/normas , Biologia Computacional/tendências , Segurança Computacional , Mineração de Dados/ética , Mineração de Dados/normas , Mineração de Dados/tendências , Privacidade Genética/ética , Privacidade Genética/legislação & jurisprudência , Privacidade Genética/normas , Privacidade Genética/tendências , Genômica/ética , Genômica/normas , Genômica/tendências , Humanos , Consentimento Livre e Esclarecido/legislação & jurisprudência , Consentimento Livre e Esclarecido/normas , Registro Médico Coordenado/normas , Gestão de Riscos , Estados Unidos
14.
IEEE J Biomed Health Inform ; 21(5): 1466-1472, 2017 09.
Artigo em Inglês | MEDLINE | ID: mdl-27834660

RESUMO

Applications of genomic studies are spreading rapidly in many domains of science and technology such as healthcare, biomedical research, direct-to-consumer services, and legal and forensic. However, there are a number of obstacles that make it hard to access and process a big genomic database for these applications. First, sequencing genomic sequence is a time consuming and expensive process. Second, it requires large-scale computation and storage systems to process genomic sequences. Third, genomic databases are often owned by different organizations, and thus, not available for public usage. Cloud computing paradigm can be leveraged to facilitate the creation and sharing of big genomic databases for these applications. Genomic data owners can outsource their databases in a centralized cloud server to ease the access of their databases. However, data owners are reluctant to adopt this model, as it requires outsourcing the data to an untrusted cloud service provider that may cause data breaches. In this paper, we propose a privacy-preserving model for outsourcing genomic data to a cloud. The proposed model enables query processing while providing privacy protection of genomic databases. Privacy of the individuals is guaranteed by permuting and adding fake genomic records in the database. These techniques allow cloud to evaluate count and top-k queries securely and efficiently. Experimental results demonstrate that a count and a top-k query over 40 Single Nucleotide Polymorphisms (SNPs) in a database of 20 000 records takes around 100 and 150 s, respectively.


Assuntos
Segurança Computacional , Genômica/normas , Informática Médica/normas , Serviços Terceirizados/normas , Privacidade , Computação em Nuvem , Bases de Dados Genéticas , Humanos
15.
BMC Med Inform Decis Mak ; 16 Suppl 3: 73, 2016 07 25.
Artigo em Inglês | MEDLINE | ID: mdl-27454233

RESUMO

BACKGROUND: Accurately assessing pain for those who cannot make self-report of pain, such as minimally responsive or severely brain-injured patients, is challenging. In this paper, we attempted to address this challenge by answering the following questions: (1) if the pain has dependency structures in electronic signals and if so, (2) how to apply this pattern in predicting the state of pain. To this end, we have been investigating and comparing the performance of several machine learning techniques. METHODS: We first adopted different strategies, in which the collected original n-dimensional numerical data were converted into binary data. Pain states are represented in binary format and bound with above binary features to construct (n + 1) -dimensional data. We then modeled the joint distribution over all variables in this data using the Restricted Boltzmann Machine (RBM). RESULTS: Seventy-eight pain data items were collected. Four individuals with the number of recorded labels larger than 1000 were used in the experiment. Number of avaliable data items for the four patients varied from 22 to 28. Discriminant RBM achieved better accuracy in all four experiments. CONCLUSION: The experimental results show that RBM models the distribution of our binary pain data well. We showed that discriminant RBM can be used in a classification task, and the initial result is advantageous over other classifiers such as support vector machine (SVM) using PCA representation and the LDA discriminant method.


Assuntos
Dor/diagnóstico , Reconhecimento Automatizado de Padrão , Humanos , Redes Neurais de Computação
16.
BMC Med Inform Decis Mak ; 15 Suppl 5: S5, 2015.
Artigo em Inglês | MEDLINE | ID: mdl-26733391

RESUMO

BACKGROUND: The increasing availability of genome data motivates massive research studies in personalized treatment and precision medicine. Public cloud services provide a flexible way to mitigate the storage and computation burden in conducting genome-wide association studies (GWAS). However, data privacy has been widely concerned when sharing the sensitive information in a cloud environment. METHODS: We presented a novel framework (FORESEE: Fully Outsourced secuRe gEnome Study basEd on homomorphic Encryption) to fully outsource GWAS (i.e., chi-square statistic computation) using homomorphic encryption. The proposed framework enables secure divisions over encrypted data. We introduced two division protocols (i.e., secure errorless division and secure approximation division) with a trade-off between complexity and accuracy in computing chi-square statistics. RESULTS: The proposed framework was evaluated for the task of chi-square statistic computation with two case-control datasets from the 2015 iDASH genome privacy protection challenge. Experimental results show that the performance of FORESEE can be significantly improved through algorithmic optimization and parallel computation. Remarkably, the secure approximation division provides significant performance gain, but without missing any significance SNPs in the chi-square association test using the aforementioned datasets. CONCLUSIONS: Unlike many existing HME based studies, in which final results need to be computed by the data owner due to the lack of the secure division operation, the proposed FORESEE framework support complete outsourcing to the cloud and output the final encrypted chi-square statistics.


Assuntos
Computação em Nuvem/normas , Segurança Computacional/normas , Privacidade Genética/normas , Estudo de Associação Genômica Ampla/normas , Humanos , Serviços Terceirizados/normas
17.
BMC Med Inform Decis Mak ; 14 Suppl 1: S1, 2014.
Artigo em Inglês | MEDLINE | ID: mdl-25521230

RESUMO

To answer the need for the rigorous protection of biomedical data, we organized the Critical Assessment of Data Privacy and Protection initiative as a community effort to evaluate privacy-preserving dissemination techniques for biomedical data. We focused on the challenge of sharing aggregate human genomic data (e.g., allele frequencies) in a way that preserves the privacy of the data donors, without undermining the utility of genome-wide association studies (GWAS) or impeding their dissemination. Specifically, we designed two problems for disseminating the raw data and the analysis outcome, respectively, based on publicly available data from HapMap and from the Personal Genome Project. A total of six teams participated in the challenges. The final results were presented at a workshop of the iDASH (integrating Data for Analysis, 'anonymization,' and SHaring) National Center for Biomedical Computing. We report the results of the challenge and our findings about the current genome privacy protection techniques.


Assuntos
Privacidade Genética/normas , Estudo de Associação Genômica Ampla/normas , Disseminação de Informação , Humanos
18.
Stud Health Technol Inform ; 201: 433-40, 2014.
Artigo em Inglês | MEDLINE | ID: mdl-24943578

RESUMO

Accurate assessment and documentation of skin conditions facilitate communication among care providers and are critical to effective prevention and mitigation of pressure ulcer. We report developing a prototype mobile system called SAPPIRE (Skin Assessment for Pressure Ulcer Prevention, an Integrated Recording Environment) for an android device to assist nurses with skin assessment and documentation at bedside. SAPPIRE demonstrates (1) data documentation conforming to the relevant terminology standards, (2) data exchange using Continuity of Care Records (CCR) standard and (3) smart display of patient data relevant to risk parameters to promote accurate pressure ulcer risk assessment with the Braden scale. Challenges associated standardizing assessment data faced during this development and the approaches that SAPPIRE took to overcome them are described.


Assuntos
Registros Eletrônicos de Saúde/organização & administração , Diagnóstico de Enfermagem/métodos , Registros de Enfermagem , Úlcera por Pressão/diagnóstico , Úlcera por Pressão/enfermagem , Software , Sistemas de Apoio a Decisões Clínicas/organização & administração , Controle de Formulários e Registros , Humanos , Armazenamento e Recuperação da Informação/métodos , Sistemas Automatizados de Assistência Junto ao Leito , Medição de Risco , Telemedicina/métodos
19.
IEEE Trans Circuits Syst Video Technol ; 23(11): 1941-1956, 2013 Nov.
Artigo em Inglês | MEDLINE | ID: mdl-25505829

RESUMO

This paper proposes a novel model on intra coding for High Efficiency Video Coding (HEVC), which simultaneously predicts blocks of pixels with optimal rate distortion. It utilizes the spatial statistical correlation for the optimal prediction based on 2-D contexts, in addition to formulating the data-driven structural interdependences to make the prediction error coherent with the probability distribution, which is desirable for successful transform and coding. The structured set prediction model incorporates a max-margin Markov network (M3N) to regulate and optimize multiple block predictions. The model parameters are learned by discriminating the actual pixel value from other possible estimates to maximize the margin (i.e., decision boundary bandwidth). Compared to existing methods that focus on minimizing prediction error, the M3N-based model adaptively maintains the coherence for a set of predictions. Specifically, the proposed model concurrently optimizes a set of predictions by associating the loss for individual blocks to the joint distribution of succeeding discrete cosine transform coefficients. When the sample size grows, the prediction error is asymptotically upper bounded by the training error under the decomposable loss function. As an internal step, we optimize the underlying Markov network structure to find states that achieve the maximal energy using expectation propagation. For validation, we integrate the proposed model into HEVC for optimal mode selection on rate-distortion optimization. The proposed prediction model obtains up to 2.85% bit rate reduction and achieves better visual quality in comparison to the HEVC intra coding.

20.
J Am Med Inform Assoc ; 19(2): 196-201, 2012.
Artigo em Inglês | MEDLINE | ID: mdl-22081224

RESUMO

iDASH (integrating data for analysis, anonymization, and sharing) is the newest National Center for Biomedical Computing funded by the NIH. It focuses on algorithms and tools for sharing data in a privacy-preserving manner. Foundational privacy technology research performed within iDASH is coupled with innovative engineering for collaborative tool development and data-sharing capabilities in a private Health Insurance Portability and Accountability Act (HIPAA)-certified cloud. Driving Biological Projects, which span different biological levels (from molecules to individuals to populations) and focus on various health conditions, help guide research and development within this Center. Furthermore, training and dissemination efforts connect the Center with its stakeholders and educate data owners and data consumers on how to share and use clinical and biological data. Through these various mechanisms, iDASH implements its goal of providing biomedical and behavioral researchers with access to data, software, and a high-performance computing environment, thus enabling them to generate and test new hypotheses.


Assuntos
Algoritmos , Confidencialidade , Disseminação de Informação , Informática Médica , Previsões , Objetivos , Health Insurance Portability and Accountability Act , Armazenamento e Recuperação da Informação , Estados Unidos
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA