Your browser doesn't support javascript.
loading
Manual Evaluation of Record Linkage Algorithm Performance in Four Real-World Datasets.
Gupta, Agrayan K; Xu, Huiping; Li, Xiaochun; Vest, Joshua R; Grannis, Shaun J.
Afiliación
  • Gupta AK; Indiana University School of Medicine, Indianapolis, Indiana, United States.
  • Xu H; Department of Biostatistics, Indiana University School of Medicine, Indianapolis, Indiana, United States.
  • Li X; Department of Biostatistics, Indiana University School of Medicine, Indianapolis, Indiana, United States.
  • Vest JR; Center for Biomedical Informatics, Regenstrief Institute, Indianapolis, Indiana, United States.
  • Grannis SJ; Department of Health Policy and Management, Indiana University Richard M. Fairbanks School of Public Health, Indianapolis, Indiana, United States.
Appl Clin Inform ; 15(3): 620-628, 2024 May.
Article en En | MEDLINE | ID: mdl-38508580
ABSTRACT

OBJECTIVES:

Patient data are fragmented across multiple repositories, yielding suboptimal and costly care. Record linkage algorithms are widely accepted solutions for improving completeness of patient records. However, studies often fail to fully describe their linkage techniques. Further, while many frameworks evaluate record linkage methods, few focus on producing gold standard datasets. This highlights a need to assess these frameworks and their real-world performance. We use real-world datasets and expand upon previous frameworks to evaluate a consistent approach to the manual review of gold standard datasets and measure its impact on algorithm performance.

METHODS:

We applied the framework, which includes elements for data description, reviewer training and adjudication, and software and reviewer descriptions, to four datasets. Record pairs were formed and between 15,000 and 16,500 records were randomly sampled from these pairs. After training, two reviewers determined match status for each record pair. If reviewers disagreed, a third reviewer was used for final adjudication.

RESULTS:

Between the four datasets, the percent discordant rate ranged from 1.8 to 13.6%. While reviewers' discordance rate typically ranged between 1 and 5%, one exhibited a 59% discordance rate, showing the importance of the third reviewer. The original analysis was compared with three sensitivity analyses. The original analysis most often exhibited the highest predictive values compared with the sensitivity analyses.

CONCLUSION:

Reviewers vary in their assessment of a gold standard, which can lead to variances in estimates for matching performance. Our analysis demonstrates how a multireviewer process can be applied to create gold standards, identify reviewer discrepancies, and evaluate algorithm performance.
Asunto(s)

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Asunto principal: Algoritmos Límite: Humans Idioma: En Revista: Appl Clin Inform Año: 2024 Tipo del documento: Article País de afiliación: Estados Unidos

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Asunto principal: Algoritmos Límite: Humans Idioma: En Revista: Appl Clin Inform Año: 2024 Tipo del documento: Article País de afiliación: Estados Unidos