Appl Clin Inform
DOI: 10.1055/a-2291-1391
Research Article

Manual evaluation of record linkage algorithm performance in four real-world datasets

Agrayan Gupta
1   Regenstrief Institute Inc, Indianapolis, United States (Ringgold ID: RIN50826)
2   Indiana University School of Medicine, Indianapolis, United States (Ringgold ID: RIN12250)
,
Huiping Xu
3   Department of Biostatistics, Indiana University School of Medicine, Indianapolis, United States (Ringgold ID: RIN12250)
,
Xiaochun Li
4   Department of Biostatistics, Regenstrief Institute Inc, Indianapolis, United States (Ringgold ID: RIN50826)
,
Josh Vest
5   Center for Biomedical Informatics, Regenstrief Institute Inc, Indianapolis, United States (Ringgold ID: RIN50826)
6   Department of Health Policy & Management, Indiana University Richard M Fairbanks School of Public Health, Indianapolis, United States (Ringgold ID: RIN124006)
,
Shaun Grannis
7   Regenstrief Institute Inc, Indianapolis, United States (Ringgold ID: RIN50826)
8   Indiana University School of Medicine, Indianapolis, United States (Ringgold ID: RIN12250)
› Author Affiliations
Supported by: Patient-Centered Outcomes Research Institute ME-2017C1-6425

Background: Patient data is fragmented across multiple repositories, yielding suboptimal and costly care. Record linkage algorithms are widely accepted solutions for improving completeness of patient records. However, studies often fail to fully describe their linkage techniques. Further, while many frameworks evaluate record linkage methods, few focus on producing gold standard datasets. This highlights a need to assess these frameworks and their real-world performance. Objective: We use real-world datasets and expand upon previous frameworks to evaluate a consistent approach to the manual review of gold standard datasets and measure its impact on algorithm performance. Methods: We applied the framework, which includes elements for data description, reviewer training and adjudication, and software and reviewer descriptions, to four datasets. Record-pairs were formed and 15,000 records were randomly sampled from these pairs. After training, two reviewers determined match status for each record-pair. If reviewers disagreed, a third reviewer was used for final adjudication. Results: Between the four datasets, the percent discordant rate ranged from 1.8-13.6%. While reviewers’ discordance rate typically ranged between 1% and 5%, one exhibited a 59% discordance rate, showing the importance of the third reviewer. The original analysis was compared to three sensitivity analyses. The original analysis most often exhibited the highest predictive values compared to the sensitivity analyses. Conclusion: Reviewers vary in their assessment of a gold standard, which can lead to variances in estimates for matching performance. Our analysis demonstrates how a multi-reviewer process can be applied to create gold standards, identify reviewer discrepancies, and evaluate algorithm performance.



Publication History

Received: 26 September 2023

Accepted after revision: 18 March 2024

Accepted Manuscript online:
20 March 2024

© . Thieme. All rights reserved.

Georg Thieme Verlag KG
Rüdigerstraße 14, 70469 Stuttgart, Germany