Making work visible for electronic phenotype implementation: Lessons learned from the eMERGE network

Ning Shang, Cong Liu, Luke V. Rasmussen, Casey N. Ta, Robert J. Caroll, Barbara Benoit, Todd Lingren, Ozan Dikilitas, Frank D. Mentch, David S. Carrell, Wei Qi Wei, Yuan Luo, Vivian S. Gainer, Iftikhar J. Kullo, Jennifer A. Pacheco, Hakon Hakonarson, Theresa L. Walunas, Joshua C. Denny, Ken Wiley, Shawn N. MurphyGeorge Hripcsak*, Chunhua Weng

*Corresponding author for this work

Research output: Contribution to journalArticle

Abstract

Background: Implementation of phenotype algorithms requires phenotype engineers to interpret human-readable algorithms and translate the description (text and flowcharts) into computable phenotypes – a process that can be labor intensive and error prone. To address the critical need for reducing the implementation efforts, it is important to develop portable algorithms. Methods: We conducted a retrospective analysis of phenotype algorithms developed in the Electronic Medical Records and Genomics (eMERGE) network and identified common customization tasks required for implementation. A novel scoring system was developed to quantify portability from three aspects: Knowledge conversion, clause Interpretation, and Programming (KIP). Tasks were grouped into twenty representative categories. Experienced phenotype engineers were asked to estimate the average time spent on each category and evaluate time saving enabled by a common data model (CDM), specifically the Observational Medical Outcomes Partnership (OMOP) model, for each category. Results: A total of 485 distinct clauses (phenotype criteria) were identified from 55 phenotype algorithms, corresponding to 1153 customization tasks. In addition to 25 non-phenotype-specific tasks, 46 tasks are related to interpretation, 613 tasks are related to knowledge conversion, and 469 tasks are related to programming. A score between 0 and 2 (0 for easy, 1 for moderate, and 2 for difficult portability) is assigned for each aspect, yielding a total KIP score range of 0 to 6. The average clause-wise KIP score to reflect portability is 1.37 ± 1.38. Specifically, the average knowledge (K) score is 0.64 ± 0.66, interpretation (I) score is 0.33 ± 0.55, and programming (P) score is 0.40 ± 0.64. 5% of the categories can be completed within one hour (median). 70% of the categories take from days to months to complete. The OMOP model can assist with vocabulary mapping tasks. Conclusion: This study presents firsthand knowledge of the substantial implementation efforts in phenotyping and introduces a novel metric (KIP) to measure portability of phenotype algorithms for quantifying such efforts across the eMERGE Network. Phenotype developers are encouraged to analyze and optimize the portability in regards to knowledge, interpretation and programming. CDMs can be used to improve the portability for some ‘knowledge-oriented’ tasks.

Original languageEnglish (US)
Article number103293
JournalJournal of Biomedical Informatics
Volume99
DOIs
StatePublished - Nov 2019

    Fingerprint

Keywords

  • Electronic health records
  • Phenotyping
  • Portability

ASJC Scopus subject areas

  • Computer Science Applications
  • Health Informatics

Cite this

Shang, N., Liu, C., Rasmussen, L. V., Ta, C. N., Caroll, R. J., Benoit, B., Lingren, T., Dikilitas, O., Mentch, F. D., Carrell, D. S., Wei, W. Q., Luo, Y., Gainer, V. S., Kullo, I. J., Pacheco, J. A., Hakonarson, H., Walunas, T. L., Denny, J. C., Wiley, K., ... Weng, C. (2019). Making work visible for electronic phenotype implementation: Lessons learned from the eMERGE network. Journal of Biomedical Informatics, 99, [103293]. https://doi.org/10.1016/j.jbi.2019.103293