Your browser doesn't support javascript.
loading
Lessons Learned for Identifying and Annotating Permissions in Clinical Consent Forms.
Umberfield, Elizabeth E; Jiang, Yun; Fenton, Susan H; Stansbury, Cooper; Ford, Kathleen; Crist, Kaycee; Kardia, Sharon L R; Thomer, Andrea K; Harris, Marcelline R.
Afiliação
  • Umberfield EE; Health Policy & Management, Indiana University Richard M Fairbanks School of Public Health, Indianapolis, Indiana, United States.
  • Jiang Y; Center for Biomedical Informatics, Regenstrief Institute, Inc., Indianapolis, Indiana, United States.
  • Fenton SH; Department of Systems, Populations and Leadership, University of Michigan School of Nursing, Ann Arbor, Michigan, United States.
  • Stansbury C; School of Biomedical Informatics, University of Texas Health Science Center, Houston, Texas, United States.
  • Ford K; Department of Computational Medicine and Bioinformatics, University of Michigan Medical School, Ann Arbor, Michigan, United States.
  • Crist K; The Michigan Institute for Computational Discovery and Engineering, University of Michigan, Ann Arbor, Michigan, United States.
  • Kardia SLR; Department of Systems, Populations and Leadership, University of Michigan School of Nursing, Ann Arbor, Michigan, United States.
  • Thomer AK; Rory Meyers School of Nursing, New York University, New York, New York, United States.
  • Harris MR; Department of Epidemiology, University of Michigan School of Public Health, Ann Arbor, Michigan, United States.
Appl Clin Inform ; 12(3): 429-435, 2021 05.
Article em En | MEDLINE | ID: mdl-34161986
BACKGROUND: The lack of machine-interpretable representations of consent permissions precludes development of tools that act upon permissions across information ecosystems, at scale. OBJECTIVES: To report the process, results, and lessons learned while annotating permissions in clinical consent forms. METHODS: We conducted a retrospective analysis of clinical consent forms. We developed an annotation scheme following the MAMA (Model-Annotate-Model-Annotate) cycle and evaluated interannotator agreement (IAA) using observed agreement (A o), weighted kappa (κw ), and Krippendorff's α. RESULTS: The final dataset included 6,399 sentences from 134 clinical consent forms. Complete agreement was achieved for 5,871 sentences, including 211 positively identified and 5,660 negatively identified as permission-sentences across all three annotators (A o = 0.944, Krippendorff's α = 0.599). These values reflect moderate to substantial IAA. Although permission-sentences contain a set of common words and structure, disagreements between annotators are largely explained by lexical variability and ambiguity in sentence meaning. CONCLUSION: Our findings point to the complexity of identifying permission-sentences within the clinical consent forms. We present our results in light of lessons learned, which may serve as a launching point for developing tools for automated permission extraction.
Assuntos

Texto completo: 1 Base de dados: MEDLINE Assunto principal: Termos de Consentimento Tipo de estudo: Observational_studies / Prognostic_studies Idioma: En Ano de publicação: 2021 Tipo de documento: Article

Texto completo: 1 Base de dados: MEDLINE Assunto principal: Termos de Consentimento Tipo de estudo: Observational_studies / Prognostic_studies Idioma: En Ano de publicação: 2021 Tipo de documento: Article