<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="review-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id><journal-id journal-id-type="publisher-id">medinform</journal-id><journal-id journal-id-type="index">7</journal-id><journal-title>JMIR Medical Informatics</journal-title><abbrev-journal-title>JMIR Med Inform</abbrev-journal-title><issn pub-type="epub">2291-9694</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v12i1e58130</article-id><article-id pub-id-type="doi">10.2196/58130</article-id><article-categories><subj-group subj-group-type="heading"><subject>Review</subject></subj-group></article-categories><title-group><article-title>Electronic Health Record Data Quality and Performance Assessments: Scoping Review</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Penev</surname><given-names>Yordan P</given-names></name><degrees>MTM</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Buchanan</surname><given-names>Timothy R</given-names></name><degrees>BS</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Ruppert</surname><given-names>Matthew M</given-names></name><degrees>MS</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Liu</surname><given-names>Michelle</given-names></name><degrees>BS</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Shekouhi</surname><given-names>Ramin</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Guan</surname><given-names>Ziyuan</given-names></name><degrees>MS</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Balch</surname><given-names>Jeremy</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Ozrazgat-Baslanti</surname><given-names>Tezcan</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Shickel</surname><given-names>Benjamin</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Loftus</surname><given-names>Tyler J</given-names></name><degrees>MD, PhD</degrees><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Bihorac</surname><given-names>Azra</given-names></name><degrees>MS, MD</degrees><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff5">5</xref></contrib></contrib-group><aff id="aff1"><institution>Department of Medicine, University of Florida</institution>, <addr-line>Gainesville</addr-line><addr-line>Florida</addr-line>, <country>United States</country></aff><aff id="aff2"><institution>Intelligent Clinical Care Center, University of Florida</institution>, <addr-line>Gainesville</addr-line><addr-line>Florida</addr-line>, <country>United States</country></aff><aff id="aff3"><institution>College of Medicine, University of Central Florida</institution>, <addr-line>Orlando</addr-line><addr-line>Florida</addr-line>, <country>United States</country></aff><aff id="aff4"><institution>Department of Surgery, University of Florida</institution>, <addr-line>Gainesville</addr-line><addr-line>Florida</addr-line>, <country>United States</country></aff><aff id="aff5"><institution>Department of Medicine, Division of Nephrology, Hypertension, and Renal Transplantation, University of Florida</institution>, <addr-line>PO Box 100224</addr-line><addr-line>Gainesville</addr-line><addr-line>Florida</addr-line>, <country>United States</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Lovis</surname><given-names>Christian</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Mun</surname><given-names>Michelle</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Liu</surname><given-names>Zhengyang</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Azra Bihorac, MS, MD, Department of Medicine, Division of Nephrology, Hypertension, and Renal Transplantation, University of Florida, PO Box 100224, Gainesville, Florida, 32610-0224, United States, 1 3522948580, 1 3523925365; <email>abihorac@ufl.edu</email></corresp></author-notes><pub-date pub-type="collection"><year>2024</year></pub-date><pub-date pub-type="epub"><day>6</day><month>11</month><year>2024</year></pub-date><volume>12</volume><elocation-id>e58130</elocation-id><history><date date-type="received"><day>06</day><month>03</month><year>2024</year></date><date date-type="rev-recd"><day>14</day><month>05</month><year>2024</year></date><date date-type="accepted"><day>08</day><month>06</month><year>2024</year></date></history><copyright-statement>&#x00A9; Yordan P Penev, Timothy R Buchanan, Matthew M Ruppert, Michelle Liu, Ramin Shekouhi, Ziyuan Guan, Jeremy Balch, Tezcan Ozrazgat-Baslanti, Benjamin Shickel, Tyler J Loftus, Azra Bihorac. Originally published in JMIR Medical Informatics (<ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org">https://medinform.jmir.org</ext-link>), 6.11.2024. </copyright-statement><copyright-year>2024</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org/">https://medinform.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://medinform.jmir.org/2024/1/e58130"/><abstract><sec><title>Background</title><p>Electronic health records (EHRs) have an enormous potential to advance medical research and practice through easily accessible and interpretable EHR-derived databases. Attainability of this potential is limited by issues with data quality (DQ) and performance assessment.</p></sec><sec><title>Objective</title><p>This review aims to streamline the current best practices on EHR DQ and performance assessments as a replicable standard for researchers in the field.</p></sec><sec sec-type="methods"><title>Methods</title><p>PubMed was systematically searched for original research articles assessing EHR DQ and performance from inception until May 7, 2023.</p></sec><sec sec-type="results"><title>Results</title><p>Our search yielded 26 original research articles. Most articles had 1 or more significant limitations, including incomplete or inconsistent reporting (n=6, 30%), poor replicability (n=5, 25%), and limited generalizability of results (n=5, 25%). Completeness (n=21, 81%), conformance (n=18, 69%), and plausibility (n=16, 62%) were the most cited indicators of DQ, while correctness or accuracy (n=14, 54%) was most cited for data performance, with context-specific supplementation by recency (n=7, 27%), fairness (n=6, 23%), stability (n=4, 15%), and shareability (n=2, 8%) assessments. Artificial intelligence&#x2013;based techniques, including natural language data extraction, data imputation, and fairness algorithms, were demonstrated to play a rising role in improving both dataset quality and performance.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>This review highlights the need for incentivizing DQ and performance assessments and their standardization. The results suggest the usefulness of artificial intelligence&#x2013;based techniques for enhancing DQ and performance to unlock the full potential of EHRs to improve medical research and practice.</p></sec></abstract><kwd-group><kwd>electronic health record</kwd><kwd>EHR</kwd><kwd>record</kwd><kwd>data quality</kwd><kwd>data performance</kwd><kwd>clinical informatics</kwd><kwd>performance</kwd><kwd>data science</kwd><kwd>synthesis</kwd><kwd>review methods</kwd><kwd>review methodology</kwd><kwd>search</kwd><kwd>scoping</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>The adoption of electronic health records (EHRs) optimistically promises easily searchable databases as an accessible means for prospective and retrospective research applications [<xref ref-type="bibr" rid="ref1">1</xref>]. EHRs often fall short of these promises due to limited local data and poor data quality (DQ) [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref3">3</xref>]. To overcome these limitations, several institutions have harmonized databases and model ontologies, including PCORnet (The National Patient-Centered Clinical Research Network), All of Us, MIRACUM (Medical Informatics in Research and Care in University Medicine), and the EHDEN Project [<xref ref-type="bibr" rid="ref4">4</xref>-<xref ref-type="bibr" rid="ref7">7</xref>]. These programs strive to offer high-quality data for research purposes [<xref ref-type="bibr" rid="ref2">2</xref>]. However, EHR DQ remains highly variable, with some studies showing completeness in EHR parameter values ranging from 60% to 100% [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>]. Similar inconsistencies present a significant limitation to the generalizability and applicability of lessons learned across these datasets for broader medical and research purposes.</p><p>Multiple initiatives have aimed to measure and improve EHR data [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref11">11</xref>]. Early efforts in DQ assessment (DQA) demonstrated inconsistent reporting and a need for universal terminology standards in DQA efforts [<xref ref-type="bibr" rid="ref11">11</xref>]. In response, attempts at a standardized ontology for DQA have been developed, such as through the efforts of the International Consortium for Health Outcomes Measurement, 3&#x00D7;3 DQA guidelines, and the terminologies proposed by Kahn et al [<xref ref-type="bibr" rid="ref12">12</xref>] and Wang et al [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref12">12</xref>-<xref ref-type="bibr" rid="ref15">15</xref>]. More recently, artificial intelligence (AI) and natural language processing techniques have automated quality initiatives, including data assessment and augmentation [<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref17">17</xref>]. Nonetheless, these techniques introduce their own set of quality requirements, including fairness metrics, handling intolerable or lost data, and mitigating data drift [<xref ref-type="bibr" rid="ref18">18</xref>]. Measuring the result of these techniques&#x2019; application in real-world clinical contexts has given rise to another field that has become crucial for EHR improvement, namely, data performance assessment (DPA) [<xref ref-type="bibr" rid="ref19">19</xref>].</p><p>In this review, we critically evaluate peer-reviewed literature on the intersection of DQA and DPA applications, as well as trends in their automation [<xref ref-type="bibr" rid="ref10">10</xref>-<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref20">20</xref>-<xref ref-type="bibr" rid="ref22">22</xref>]. The purpose of this scoping review was to combine the 3 to formulate a more clear road map for evaluating EHR datasets for medical research and practice.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Overview</title><p>This scoping literature review was conducted according to the 2018 PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews), whose checklist is shown in <xref ref-type="supplementary-material" rid="app3">Checklist 1</xref> [<xref ref-type="bibr" rid="ref23">23</xref>].</p></sec><sec id="s2-2"><title>Literature Search</title><p>A search was performed for all full-text research articles published in English in PubMed from inception to May 7, 2023. A list of the exact search terms is included in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p></sec><sec id="s2-3"><title>Article Selection</title><p>Four investigators (JB, RS, TRB, and YPP) reviewed the selected studies during the title and abstract screening. Further 4 investigators (ML, RS, TOB, and YPP) conducted the full-text review and final extraction of articles. Title or abstract screening, full-text review, and final extraction were based on the consensus opinion between 2 independent reviewers. Conflicts were resolved by a third reviewer. Article management and calculations of interrater reliability and Cohen &#x03BA; were performed using Covidence systematic review software (Veritas Health Innovation).</p></sec><sec id="s2-4"><title>Inclusion Criteria</title><p>Titles and abstracts were screened to include original research articles assessing the DQ and performance of all or part of a hospital&#x2019;s EHR system. We looked for studies reporting on 1 or more aspects of DQ (the assessment of EHR data without consideration of follow-up actions) and data performance (the assessment of EHR data applications) as defined (<xref ref-type="table" rid="table1">Table 1</xref>).</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Data quality and performance indicator definitions, mitigation strategies, and references.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom"/><td align="left" valign="bottom">Definition</td><td align="left" valign="bottom">Mitigation strategies</td><td align="left" valign="bottom">Relevant studies</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="4"><bold>Data quality</bold></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Completeness (or, conversely, missingness)</td><td align="left" valign="top">The absence of data points, without reference to data type or plausibility [<xref ref-type="bibr" rid="ref12">12</xref>]</td><td align="left" valign="top">Automated data extraction; data imputation</td><td align="char" char="." valign="top">[<xref ref-type="bibr" rid="ref2">2</xref>-<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref24">24</xref>-<xref ref-type="bibr" rid="ref37">37</xref>]</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Conformance</td><td align="left" valign="top">The compliance of data with expected formatting, relational, or absolute definitions [<xref ref-type="bibr" rid="ref12">12</xref>]</td><td align="left" valign="top">Preemptively enforced data format standardization</td><td align="char" char="." valign="top">[<xref ref-type="bibr" rid="ref2">2</xref>-<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref24">24</xref>-<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref29">29</xref>-<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref36">36</xref>,<xref ref-type="bibr" rid="ref38">38</xref>]</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Plausibility</td><td align="left" valign="top">The possibility that a value is true given the context of other variables or temporal sequences (ie, patient date of birth must precede date of treatment or diagnosis) [<xref ref-type="bibr" rid="ref12">12</xref>]</td><td align="left" valign="top">Periodic realignment with logic rule sets or objective truth standards; thresholding</td><td align="char" char="." valign="top">[<xref ref-type="bibr" rid="ref4">4</xref>-<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref25">25</xref>,<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref28">28</xref>,<xref ref-type="bibr" rid="ref30">30</xref>-<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref35">35</xref>,<xref ref-type="bibr" rid="ref37">37</xref>-<xref ref-type="bibr" rid="ref39">39</xref>]</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Uniqueness</td><td align="left" valign="top">The lack of duplicate data among other patient records [<xref ref-type="bibr" rid="ref8">8</xref>]</td><td align="left" valign="top">Two-level encounter or visit data structure</td><td align="char" char="." valign="top">[<xref ref-type="bibr" rid="ref8">8</xref>]</td></tr><tr><td align="left" valign="top" colspan="4"><bold>Data performance</bold></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Correctness or accuracy</td><td align="left" valign="top">Whether patient records are free from errors or inconsistencies when the information provided in them is true [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref13">13</xref>]</td><td align="left" valign="top">Periodic validation against internal and external gold standards</td><td align="char" char="." valign="top">[<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref7">7</xref>-<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref23">23</xref>,<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref28">28</xref>]</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Currency or recency</td><td align="left" valign="top">Whether data were entered into the EHR<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup> within a clinically relevant time frame and is representative of the patient state at a given time of interest [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref13">13</xref>]</td><td align="left" valign="top">Enforcing predetermined hard and soft rule sets for timeline of data entry</td><td align="char" char="." valign="top">[<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref34">34</xref>,<xref ref-type="bibr" rid="ref36">36</xref>]</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Fairness (or, conversely, bias)</td><td align="left" valign="top">The degree to which data collection, augmentation, and application are free from unwarranted over- or underrepresentation of individual data elements or characteristics</td><td align="left" valign="top">Periodic review against a predetermined internal gold standard or bias criterion</td><td align="char" char="." valign="top">[<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref35">35</xref>]</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Stability (or, conversely, temporal variability)</td><td align="left" valign="top">Whether temporally dependent variables change according to predefined expectations [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref12">12</xref>]</td><td align="left" valign="top">Periodic measurement of data drift against a baseline standard of data distribution</td><td align="char" char="." valign="top">[<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref31">31</xref>]</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Shareability</td><td align="left" valign="top">Whether data can be shared directly, easily, and with no information loss [<xref ref-type="bibr" rid="ref3">3</xref>]</td><td align="left" valign="top">Preemptively enforced data standardization</td><td align="char" char="." valign="top">[<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref3">3</xref>]</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Robustness</td><td align="left" valign="top">The percent of patient records with tolerable (eg, inaccurate, inconsistent, and outdated information) versus intolerable (eg, missing required information) data quality problems [<xref ref-type="bibr" rid="ref24">24</xref>]</td><td align="left" valign="top">Timely identification of critical data quality issues</td><td align="char" char="." valign="top">[<xref ref-type="bibr" rid="ref24">24</xref>]</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>EHR: electronic health record.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s2-5"><title>Data Quality</title><sec id="s2-5-1"><title>Conformance</title><p><italic>Conformance</italic> refers to the compliance of data with expected formatting, relational, or absolute definitions [<xref ref-type="bibr" rid="ref12">12</xref>].</p></sec><sec id="s2-5-2"><title>Plausibility</title><p><italic>Plausibility</italic> refers to the possibility that a value is true given the context of other variables or temporal sequences (ie, the patient&#x2019;s date of birth must precede the date of treatment or diagnosis) [<xref ref-type="bibr" rid="ref12">12</xref>].</p></sec><sec id="s2-5-3"><title>Uniqueness</title><p><italic>Uniqueness</italic> refers to the lack of duplicated records [<xref ref-type="bibr" rid="ref8">8</xref>].</p></sec><sec id="s2-5-4"><title>Completeness (or Conversely, Missingness)</title><p>With regard to completeness, <italic>missingness</italic> is the absence of requested data points, without reference to conformance or plausibility as defined [<xref ref-type="bibr" rid="ref12">12</xref>].</p></sec></sec><sec id="s2-6"><title>Data Performance</title><sec id="s2-6-1"><title>Correctness or Accuracy</title><p><italic>Correctness</italic> or <italic>accuracy</italic> refers to whether patient records are free from errors or inconsistencies when the information provided in them is true [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref13">13</xref>].</p></sec><sec id="s2-6-2"><title>Currency or Recency</title><p><italic>Currency</italic> or <italic>recency</italic> refers to whether data were entered into the EHR within a clinically relevant time frame and are representative of the patient state at a given time of interest [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref13">13</xref>].</p></sec><sec id="s2-6-3"><title>Fairness (or Conversely, Bias)</title><p>With regard to bias, <italic>fairness</italic> refers to the degree to which data collection, augmentation, and application are free from unwarranted over- or underrepresentation of individual data elements or characteristics.</p></sec><sec id="s2-6-4"><title>Stability (or Conversely, Temporal Variability)</title><p>With regard to stability, <italic>temporal variability</italic> refers to whether temporally dependent variables change according to predefined expectations [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref12">12</xref>].</p></sec><sec id="s2-6-5"><title>Shareability</title><p><italic>Shareability</italic> refers to whether data can be shared directly, easily, and with no information loss [<xref ref-type="bibr" rid="ref3">3</xref>].</p></sec><sec id="s2-6-6"><title>Robustness</title><p><italic>Robustness</italic> refers to the percent of patient records with tolerable (eg, inaccurate, inconsistent, and outdated information) versus intolerable (eg, missing required information) DQ problems [<xref ref-type="bibr" rid="ref24">24</xref>].</p><p>We additionally included studies reporting on data imputation methods, defined as techniques used to fill in missing values in an EHR, such as through statistical approximation and the application of AI.</p></sec></sec><sec id="s2-7"><title>Exclusion Criteria</title><p>We excluded tangential analyses of DQ in articles focused primarily on clinical outcomes. As such, studies discussing data cleaning as part of quantifying clinical outcomes were excluded from our analysis. Proposals or study protocols with no results were also excluded during the screening process.</p></sec><sec id="s2-8"><title>Article Quality Assessment</title><p>Full-text articles were additionally scored as having or missing the criteria for (1) data integrity: comprehensiveness for each main outcome, including attrition and exclusions from the analysis and reasons for them; (2) method clarity: a clear description of DQA data sources, analysis steps, and criteria; (3) outcome clarity: outcomes reporting in plain language, in their entirety, and without evidence for selective reporting; and (4) generalizability: applicability of DQ techniques described in the article to other clinical settings.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Article Characteristics</title><p>The flow diagram for article selection is shown in <xref ref-type="fig" rid="figure1">Figure 1</xref>. A total of 154 records were identified using the search terms defined in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref> using the PubMed library. After the removal of 31 duplicates and the 72 articles identified as irrelevant, 51 studies proceeded to full-text review. Full-text review excluded a further 25 articles owing to reasons listed in <xref ref-type="fig" rid="figure1">Figure 1</xref>, leaving a final total of 26 original research studies [<xref ref-type="bibr" rid="ref2">2</xref>-<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref24">24</xref>-<xref ref-type="bibr" rid="ref39">39</xref>]. The Cohen &#x03BA; between the different pairs of reviewers ranged from 0.28 to 0.54 during the screening process and from 0.54 to 1.00 during the full-text review.</p><p>Study characteristics are shown in <xref ref-type="table" rid="table2">Table 2</xref> and <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>. Exactly half of the identified articles targeted general EHR DQ analysis [<xref ref-type="bibr" rid="ref4">4</xref>-<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref27">27</xref>-<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref38">38</xref>,<xref ref-type="bibr" rid="ref39">39</xref>], while the other half focused on a particular specialty or diagnosis (<xref ref-type="table" rid="table2">Table 2</xref>) [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref24">24</xref>-<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref33">33</xref>-<xref ref-type="bibr" rid="ref37">37</xref>]. The latter included primary care (n=3, 12%) [<xref ref-type="bibr" rid="ref35">35</xref>-<xref ref-type="bibr" rid="ref37">37</xref>], cardiovascular disease (n=3, 12%) [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref34">34</xref>], anesthesia or pain medicine (n=2, 8%) [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref26">26</xref>], intensive care units (n=2, 8%) [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref25">25</xref>], and pediatrics [<xref ref-type="bibr" rid="ref24">24</xref>], oncology [<xref ref-type="bibr" rid="ref2">2</xref>], and infectious disease (n=1 each, 4%) [<xref ref-type="bibr" rid="ref9">9</xref>].</p><p>Article quality assessment conducted as part of our review process identified 14 (54%) of the articles [<xref ref-type="bibr" rid="ref2">2</xref>-<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref24">24</xref>-<xref ref-type="bibr" rid="ref36">36</xref>,<xref ref-type="bibr" rid="ref38">38</xref>,<xref ref-type="bibr" rid="ref39">39</xref>] had at least 1 common study design or reporting limitation, with 5 of the articles having more than 1 [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref36">36</xref>,<xref ref-type="bibr" rid="ref38">38</xref>]. Among these, 6 (30% of all errors) articles did not clearly state their methods [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref28">28</xref>,<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref36">36</xref>,<xref ref-type="bibr" rid="ref39">39</xref>], 5 (25%) had incomplete data [<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref29">29</xref>,<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref36">36</xref>,<xref ref-type="bibr" rid="ref38">38</xref>], 5 were not generalizable to other settings [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref24">24</xref>-<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref33">33</xref>], and 4 did not clearly state their outcomes (<xref ref-type="table" rid="table2">Table 2</xref>) [<xref ref-type="bibr" rid="ref31">31</xref>,<xref ref-type="bibr" rid="ref34">34</xref>,<xref ref-type="bibr" rid="ref38">38</xref>,<xref ref-type="bibr" rid="ref39">39</xref>].</p><p>Commonly referenced DQ and performance indicators are summarized in <xref ref-type="table" rid="table3">Table 3</xref>. Respective definitions, mitigation strategies, and references are listed in <xref ref-type="table" rid="table1">Table 1</xref>.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>PRISMA 2020 flow diagram detailing study selection and reasons for exclusion for all articles considered for this scoping review. PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v12i1e58130_fig01.png"/></fig><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Frequency of clinical specialties among all papers and study limitations among all limitations identified by reviewers in this analysis.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top" colspan="2">Setting</td><td align="left" valign="top">Values, n (%)</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="3"><bold>Specialty</bold></td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">ICU<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td><td align="left" valign="top">2 (8)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Anesthesia or pain med</td><td align="left" valign="top">2 (8)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">General</td><td align="left" valign="top">13 (50)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Cardiovascular</td><td align="left" valign="top">3 (12)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Infectious disease</td><td align="left" valign="top">1 (4)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Oncology</td><td align="left" valign="top">1 (4)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Pain medicine</td><td align="left" valign="top">0 (0)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Pediatrics</td><td align="left" valign="top">1 (4)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Primary care</td><td align="left" valign="top">3 (12)</td></tr><tr><td align="left" valign="top" colspan="3"><bold>Limitations</bold></td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Incomplete data</td><td align="left" valign="top">5 (25)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Methods not clearly stated</td><td align="left" valign="top">6 (30)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Outcomes not clearly stated</td><td align="left" valign="top">4 (20)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Not generalizable to other settings</td><td align="left" valign="top">5 (25)</td></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup>ICU: intensive care unit.</p></fn></table-wrap-foot></table-wrap><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Elements of data quality and performance commonly referenced by papers included in this review.</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top" colspan="2">Data Quality and Performance Element</td><td align="left" valign="top">Values, n (%)</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="3"><bold>Data quality</bold></td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Completeness</td><td align="left" valign="top">21 (81)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Conformance</td><td align="left" valign="top">18 (69)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Plausibility</td><td align="left" valign="top">16 (62)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Uniqueness</td><td align="left" valign="top">1 (4)</td></tr><tr><td align="left" valign="top" colspan="3"><bold>Data performance</bold></td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Correctness or accuracy</td><td align="left" valign="top">14 (54)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Currency</td><td align="left" valign="top">7 (27)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Fairness or bias</td><td align="left" valign="top">6 (23)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Stability</td><td align="left" valign="top">4 (15)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Shareability</td><td align="left" valign="top">2 (8)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Robustness</td><td align="left" valign="top">1 (4)</td></tr></tbody></table></table-wrap></sec><sec id="s3-2"><title>Data Quality Assessment</title><sec id="s3-2-1"><title>Completeness</title><p>Completeness was the most cited element of DQ analysis, with references in 21 (81%) of all articles [<xref ref-type="bibr" rid="ref2">2</xref>-<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref24">24</xref>-<xref ref-type="bibr" rid="ref37">37</xref>]. Importantly, 19 (73%) studies integrated data from multiple clinical sites [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref4">4</xref>-<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref30">30</xref>-<xref ref-type="bibr" rid="ref39">39</xref>], which was associated with issues in data collection and missingness &#x201C;across organizational structure, regulation, and data sourcing&#x201D; [<xref ref-type="bibr" rid="ref31">31</xref>]. Clinical domains reported to be prone to low data completeness included patient demographics, with Estiri et al [<xref ref-type="bibr" rid="ref29">29</xref>] highlighting the issue for records of patient ethnicity and Thuraisingam et al [<xref ref-type="bibr" rid="ref35">35</xref>] for mortality records (eg, missing year of death), and medication management, with Thuraisingam et al [<xref ref-type="bibr" rid="ref35">35</xref>] highlighting the issue for dosage, strength, or frequency of prescriptions and Kiogou et al [<xref ref-type="bibr" rid="ref34">34</xref>] for missing dates or reasons for discontinuation of medications.</p><p>To combat data missingness, Lee et al [<xref ref-type="bibr" rid="ref22">22</xref>] used natural language processing algorithms to automatically extract data from patient records, while further 5 studies made use of data imputation techniques. Among the latter, 2 articles generated synthetic data, while another 3 supplemented datasets through information from external datasets. Fu et al [<xref ref-type="bibr" rid="ref3">3</xref>] generated synthetic data by modeling providers&#x2019; assessments of EHR data based on different information sources according to their individual characteristics (eg, tendency to ascertain delirium status based on Confusion Assessment Method vs prior <italic>International Statistical Classification of Diseases</italic> coding or nursing flow sheet documentation), while Zhang et al [<xref ref-type="bibr" rid="ref19">19</xref>] used a generative adversarial network (GAN) trained on real longitudinal EHR data to create single synthetic EHR episodes (eg, outpatient or inpatient visit). Meanwhile, Lee et al [<xref ref-type="bibr" rid="ref33">33</xref>] supplemented existing EHR records on heart failure by aggregating data from open-source datasets of heart failure biomarkers (including the Database of Genotypes and Phenotypes and the Biologic Specimen and Data Repository Information Coordinating Center) and using literature guidelines to create a standard set of cardiovascular outcome measures, while Curtis et al [<xref ref-type="bibr" rid="ref2">2</xref>] supplemented missing EHR mortality records with data from US Social Security Death Index and the National Death Index, and Mang et al [<xref ref-type="bibr" rid="ref30">30</xref>] used a manually generated stand-alone synthetic dataset to test the development of a new software tool for DQ assessment.</p></sec><sec id="s3-2-2"><title>Conformance</title><p>Conformance was the second most cited element of DQA, with references in 18 (69%) articles [<xref ref-type="bibr" rid="ref2">2</xref>-<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref24">24</xref>-<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref29">29</xref>-<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref36">36</xref>,<xref ref-type="bibr" rid="ref38">38</xref>]. Similar to completeness<italic>,</italic> DQ checks on conformance were performed automatically across most studies. Mitigation strategies included enforcing strict formatting rules at the time of data entry, for example, by using <italic>International Statistical Classification of Diseases</italic> codes to define the cause of death or a diagnosis of delirium [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref3">3</xref>].</p></sec><sec id="s3-2-3"><title>Plausibility</title><p>Plausibility was the third most cited element of DQA with references in 16 (62%) articles [<xref ref-type="bibr" rid="ref4">4</xref>-<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref25">25</xref>,<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref28">28</xref>,<xref ref-type="bibr" rid="ref30">30</xref>-<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref35">35</xref>,<xref ref-type="bibr" rid="ref37">37</xref>-<xref ref-type="bibr" rid="ref39">39</xref>]. Clinical domains prone to issues with plausibility included patient baseline physical characteristics, medication, and laboratory records. Estiri et al [<xref ref-type="bibr" rid="ref29">29</xref>] and Wang et al [<xref ref-type="bibr" rid="ref39">39</xref>] reported significant rates of plausibility issues for baseline physical characteristics, with higher error rates for records of patient height as compared to weight, likely due to the multiple flow sheet fields for height, including &#x201C;estimated,&#x201D; &#x201C;reported,&#x201D; and &#x201C;measured,&#x201D; which are generally averaged or selectively dropped. Pharmacologic data were prone to issues with plausibility due to timeliness (eg, antiretroviral therapy was dispensed before or more than 30 days after the visit date [<xref ref-type="bibr" rid="ref9">9</xref>]) or discrepancies between diagnoses and drugs (eg, nonsteroidal anti-inflammatory drug prescription on the date of gastroduodenal ulcer diagnosis [<xref ref-type="bibr" rid="ref6">6</xref>]). Finally, laboratory results were also prone to issues with plausibility due to value ranges, units, timing (eg, laboratory time was at an invalid time of day or in the future), and discrepancies between diagnoses and laboratory records (eg, drug was documented as present but there was no laboratory record) or drug prescriptions and laboratory records (eg, metformin was prescribed prior to a documented hemoglobin A<sub>1c </sub>laboratory result, or warfarin was prescribed without a follow-up international normalized ratio laboratory result) [<xref ref-type="bibr" rid="ref6">6</xref>]. Notably, this may reflect poorly integrated health care systems where laboratories are being drawn at disparate institutions.</p><p>A total of 18 (69%) studies used logic statements to assess plausibility [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref4">4</xref>-<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref28">28</xref>,<xref ref-type="bibr" rid="ref31">31</xref>-<xref ref-type="bibr" rid="ref38">38</xref>], including rules to determine temporal plausibility (eg, laboratories drawn at an invalid time of day [eg, 10:65 AM] [<xref ref-type="bibr" rid="ref6">6</xref>], extubation occurring prior to intubation [<xref ref-type="bibr" rid="ref14">14</xref>], or death date occurring before birth date [<xref ref-type="bibr" rid="ref32">32</xref>]), diagnostic or procedural plausibility (eg, a procedure marked as an outpatient when it is only performed on an inpatient basis [<xref ref-type="bibr" rid="ref38">38</xref>] or an obstetric diagnosis given for a biologically male patient [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref38">38</xref>]), alignment with external standards or expectations (eg, laboratory result absent for diagnosis or drug [<xref ref-type="bibr" rid="ref6">6</xref>] or demographic alignment of medication name and dose with expected value ranges [<xref ref-type="bibr" rid="ref34">34</xref>]), and others. A total of 11 (42%) studies used thresholding to identify data of low or questionable quality [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref28">28</xref>,<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref35">35</xref>,<xref ref-type="bibr" rid="ref37">37</xref>,<xref ref-type="bibr" rid="ref39">39</xref>], including clinical and physiological value ranges (eg, BMI between 12 and 90 kg/m<sup>2</sup> [<xref ref-type="bibr" rid="ref35">35</xref>] or fraction of inspired oxygen between 10% and 100% [<xref ref-type="bibr" rid="ref14">14</xref>]) and logical thresholds (eg, recorded date of arrival prior to the date of data collection initiation [<xref ref-type="bibr" rid="ref8">8</xref>] or difference of &#x003E;730 days when comparing age in years and date of birth fields [<xref ref-type="bibr" rid="ref9">9</xref>]).</p></sec><sec id="s3-2-4"><title>Uniqueness</title><p>Finally, 1 (4%) study reported on data uniqueness. Aerts et al [<xref ref-type="bibr" rid="ref8">8</xref>] measured the frequency of patient record duplications (ie, when patient records were erroneously copied during data merging or reprocessing). To reduce the rate of record duplications, the researchers in the study suggest a 2-level data structure, with more general patient data being recorded at the encounter level (which can include multiple visits during a single clinical episode) and diagnosis or procedure-specific data at the level of the particular visit.</p></sec></sec><sec id="s3-3"><title>Data Performance Assessment</title><sec id="s3-3-1"><title>Correctness or Accuracy</title><p>Correctness or accuracy was the most cited element in data performance analysis, with references in 14 (54%) of all articles [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref25">25</xref>,<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref32">32</xref>-<xref ref-type="bibr" rid="ref37">37</xref>,<xref ref-type="bibr" rid="ref39">39</xref>]. The metric was evaluated via manual review in 8 (57%) out of the 14 articles that reported the measure [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref25">25</xref>,<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref34">34</xref>,<xref ref-type="bibr" rid="ref36">36</xref>,<xref ref-type="bibr" rid="ref39">39</xref>]. A total of 5 (36%) articles evaluated it in comparison to an external standard, including national registries [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref35">35</xref>], EHR case definitions based on billing codes [<xref ref-type="bibr" rid="ref36">36</xref>], and literature guidelines with high research use [<xref ref-type="bibr" rid="ref33">33</xref>], or, in the case of a newly proposed AI technique for synthetic data augmentation, comparison to a previously published GAN model performance [<xref ref-type="bibr" rid="ref19">19</xref>]. A further 3 (21%) assessed correctness or accuracy against an internal standard by calculating the proportion of records satisfying internally predetermined rule sets [<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref37">37</xref>]. Of note, Curtis et al [<xref ref-type="bibr" rid="ref2">2</xref>] and Terry et al [<xref ref-type="bibr" rid="ref36">36</xref>] used both manual review and comparison to an external gold standard for validation.</p></sec><sec id="s3-3-2"><title>Currency or Recency</title><p>Recency was the second most cited data performance element, with references in 7 (27%) articles [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref34">34</xref>,<xref ref-type="bibr" rid="ref36">36</xref>]. Among these, 5 (71%) studies evaluated the metric according to internally predetermined hard rule sets (eg, whether a patient who is obese had a weight recording within 1 year of the previous data point or whether data were entered into the EHR within 3 days of the clinical encounter [<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref36">36</xref>]) or soft rule sets (eg, whether the data were entered into the EHR within a subjectively determined clinically actionable time limit [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref34">34</xref>]), while 2 (29%) used external standards, including national registries and guidelines [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref27">27</xref>].</p></sec><sec id="s3-3-3"><title>Fairness or Bias</title><p>The third most cited data performance element was fairness or bias, with references in 6 (23%) articles [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref35">35</xref>]. Among these, Lee et al [<xref ref-type="bibr" rid="ref22">22</xref>], Thuraisingam et al [<xref ref-type="bibr" rid="ref35">35</xref>], Tian et al [<xref ref-type="bibr" rid="ref27">27</xref>], and Garc&#x00ED;a-de-Le&#x00F3;n-Chocano et al [<xref ref-type="bibr" rid="ref24">24</xref>] assessed fairness by manual review, while Fu et al [<xref ref-type="bibr" rid="ref3">3</xref>] and Zhang et al [<xref ref-type="bibr" rid="ref19">19</xref>] did so through automated review against a predetermined internal gold standard (ie, distribution of data characteristics within a real EHR dataset) or data bias criterion (ie, critic model measuring Jensen-Shannon divergence between real and synthetic data over time), respectively.</p></sec><sec id="s3-3-4"><title>Stability</title><p>Data stability was the fourth most cited performance element, referenced in 4 (15%) articles [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref31">31</xref>]. All 4 articles that measured data stability did so via temporal statistical analyses of data drift according to a predetermined internal baseline standard of data distribution [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref37">37</xref>].</p></sec><sec id="s3-3-5"><title>Shareability</title><p>Shareability was referenced in 2 (8%) articles from our analysis [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref3">3</xref>]. Both studies measured the performance metric by way of manual review in a pre- and posttest analysis of data standardization [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref3">3</xref>].</p></sec><sec id="s3-3-6"><title>Robustness</title><p>Finally, Garc&#x00ED;a-de-Le&#x00F3;n-Chocano et al [<xref ref-type="bibr" rid="ref24">24</xref>] reported on information robustness by way of statistical estimation of critical (eg, missing or null required values) versus noncritical (all other) DQ issues that may obstruct subsequent data applications and performance measures.</p></sec></sec><sec id="s3-4"><title>Interventions for Improving DQ and Performance</title><p>Three articles included in our analysis reported effective interventions to improve DQ and performance [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref37">37</xref>]. In terms of DQ, Walker et al [<xref ref-type="bibr" rid="ref37">37</xref>] reported an increase in compliance, with 155 completeness and plausibility data checks from 53% to 100% across 6 clinical sites after 3 rounds of DQA. In terms of DQ and performance, Puttkamer et al [<xref ref-type="bibr" rid="ref9">9</xref>] reported both higher data completeness and recency following a continuous data reporting and feedback system implementation. Finally, Engel et al [<xref ref-type="bibr" rid="ref4">4</xref>] reported increased shareability (concept success rate, ie, whether data partners converted information from their individual EHRs to the shared database)&#x2014;an increase from 90% to 98.5%&#x2014;and a notable reduction in the percentage of sites with over 3 DQ errors&#x2014;a reduction from 67% to 35%&#x2014;across 50+ clinical sites over 2 years.</p></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Contributions and Comparison With Prior Work</title><p>This scoping review provides an overview of the most common and successful means of EHR DQ and performance analysis. The review adds to a growing body of literature on the subject, most recently supplemented by a systematic review by Lewis et al [<xref ref-type="bibr" rid="ref40">40</xref>]. To our knowledge, ours is the first review of specialty-specific applications of DQ alongside performance assessments. We identified and analyzed a total of 26 original research articles recently published on the topic. The results serve to characterize the most common medical fields making use of such assessments, the methodologies they use for conducting them, and areas for specialty-specific, as well as generalizable, future improvement. Finally, the discussion proposes a set of 6 unique and practical recommendations for minimizing modifiable DQ and performance issues arising during data extraction and mapping.</p></sec><sec id="s4-2"><title>Article Characteristics</title><p>Our review noted a paucity of DQ assessments within clinical specialties, where expert domain knowledge plays a key role in identifying logic inconsistencies. Half of all identified articles concerned general EHR data assessments, while the other half focused on medical fields such as primary care, cardiovascular diseases, or intensive care unit or anesthesia, with the notable absence of psychiatry, emergency medicine, and any of the surgical specialties. This points to a lack of peer-reviewed research and underuse of DQ and performance strategies across a wide spectrum of the medical field. There is a wide knowledge gap between how data are entered and acted upon clinically and how they appear in silico. Therefore, more efforts need to be directed toward supporting EHR data assessment initiatives in these specialties, with close collaboration between clinical users and data scientists.</p><p>More than half of the articles included in this scoping review had common limitations, including using or reporting incomplete data, methods, and outcomes. Among the articles scoring high for incomplete data, the chief issues include data attrition during extraction [<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref29">29</xref>] and unclear or missing reporting [<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref36">36</xref>,<xref ref-type="bibr" rid="ref38">38</xref>], pointing to a need for higher information interoperability and reporting standards, such as those put forth by Kahn et al [<xref ref-type="bibr" rid="ref12">12</xref>]. These standards recommend using a harmonized and inclusive framework for the reporting of DQ assessments, including standardized definitions for completeness, conformance, plausibility, and other measures as discussed previously.</p><p>Similar issues were observed with methods reporting, with several articles underreporting steps in their data extraction or analysis, thereby limiting the replicability and generalizability of their findings [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref28">28</xref>,<xref ref-type="bibr" rid="ref33">33</xref>]. Unclear reporting or underreporting was a substantial issue for outcomes as well, with low-scoring articles reporting only partial or too high-level results suggesting selective reporting bias [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref31">31</xref>,<xref ref-type="bibr" rid="ref34">34</xref>,<xref ref-type="bibr" rid="ref38">38</xref>]. To align with the standards set forth by articles scoring high in reporting quality, we recommend stating all data sourcing, methods, and results according to predetermined definitions of DQ or performance (see above) in enough detail such that they would be easily replicated by researchers at an unrelated institution.</p><p>A final article quality pitfall concerned articles that were too specific to a particular health system or clinical context. The chief issues among original research articles that in house scored &#x201C;low&#x201D; in our generalizability assessment concerned their overreliance on internal DQ checks or measures that could only be implemented within their specific institutional EHR [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref24">24</xref>-<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref33">33</xref>]. To increase generalizability, we recommend relying on external DQ standards such as societal guidelines, previously published measures, or open-source databases, to the extent possible before resorting to the development of new in-house tools that impose limitations to generalizability outside the local clinical context [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref12">12</xref>-<xref ref-type="bibr" rid="ref15">15</xref>].</p></sec><sec id="s4-3"><title>Data Quality Assessment</title><p>The marked drop-off between the use of completeness, conformance, and plausibility versus other indicators (<xref ref-type="table" rid="table3">Table 3</xref>) demonstrates that the field has settled on these measures as the main components of EHR DQ analysis. Taking this into consideration, we recommend measuring all 3 for a general assessment of clinical DQ. Of note, there is a significant drop-off between 81% (n=21) of studies reporting on completeness versus 69% (n=18) on conformance and 62% (n=16) on plausibility, which indicates an opportunity for limited but quick DQ &#x201C;checks&#x201D; using completeness measures only. More specialized analyses may require further reporting, including uniqueness in the event of data merger with the possibility of duplicate results. These may be particularly important in the case of EHR DQ assessments following information reconciliation from the merger of multiple data sources, including patient demographics or baseline physical characteristics and laboratory or pharmacological data, which were shown to be particularly prone to errors in DQ.</p><p>Our review additionally demonstrates that issues with data completeness, conformance, and plausibility may be at least partially addressed with data imputation methods. While previously these methods were either too limited in scope (completeness only), crude (eg, augmenting missing data with the mean of the entire dataset or a value&#x2019;s k-nearest neighbor), or computationally expensive (eg, individual values calculated via regression models based on predetermined sets of correlated features), our review suggests that these tasks are being increasingly automated. Specifically, data attrition contributing to missingness and conformity at the extraction stage may be minimized with AI data extractor algorithms, such as the one described by Lee et al [<xref ref-type="bibr" rid="ref22">22</xref>]. In cases where further extraction is no longer feasible, the dataset may be augmented by (1) using large language models for extracting structured data available in other formats (eg, laboratory values recorded in the text of media files from outside patient records); (2) incorporating or cross-referencing data from well-established outside data repositories (eg, the US Social Security Death Index for mortality records [<xref ref-type="bibr" rid="ref2">2</xref>] or the Database of Genotypes and Phenotypes and the Biologic Specimen for biomarkers of heart failure and other conditions [<xref ref-type="bibr" rid="ref33">33</xref>]); or (3) generating synthetic data, for example, by modeling providers&#x2019; behaviors with respect to different information types or sources [<xref ref-type="bibr" rid="ref3">3</xref>] and by using GANs to create synthetic care episodes based on longitudinal EHR observations [<xref ref-type="bibr" rid="ref19">19</xref>].</p></sec><sec id="s4-4"><title>Data Performance Assessment</title><p>Correctness or accuracy was by far the most reported measure among the data performance indicators examined in our review. While certainly integral to assessing a dataset&#x2019;s usability and potential for downstream clinical or research impact, correctness alone is insufficient to guarantee the success of said applications. A technically &#x201C;correct&#x201D; dataset may still be practically limited if it is outdated, biased, inconsistent, or entirely idiosyncratic. We, therefore, recommend that future data assessments consider including additional measures of recency, fairness, stability, and shareability, respectively, among their core set of performance indicators as they each contribute a unique measure of a dataset&#x2019;s applicability. Importantly, our review noted considerable heterogeneity in the definitions used for these additional measures (eg, by defining data recency in terms of whether the information was logged into the EHR within a set time or whether it represents a patient&#x2019;s state at a given time period [<xref ref-type="table" rid="table1">Table 1</xref>] [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref13">13</xref>]), suggesting that further efforts are needed to harmonize outcome definitions in the field of data performance analysis in particular. Nonetheless, the predominance of internal standard comparisons for measuring recency and stability in our review demonstrates that these indicators may be essential for individualized EHR DPAs and should, therefore, be considered on a case-by-case basis (eg, in epidemiology where the timing and consistency of reporting can be of essential importance, or quality improvement initiatives where a researcher might want to compare pre- vs postintervention results). Likewise, shareability ought to be considered in the case of assessing dataset performance for interoperability purposes (eg, with data integrations, sharing, and reporting).</p><p>As discussed previously, data fairness assessments can and should be considered for monitoring overall EHR bias, as well as the bias inherent to any data imputation methods as discussed above. Our review points to the fact that this is a rapidly developing field, with fairness assessments to date mostly requiring manual review against national guidelines or disease registries, or, in the case of synthetic data, real EHR datasets [<xref ref-type="bibr" rid="ref41">41</xref>-<xref ref-type="bibr" rid="ref43">43</xref>]. Nonetheless, such gold standards are not always readily available (eg, What is the standard distribution of age or race in the real world?), so tech-savvy researchers have more recently resorted to detecting fairness during the validation of machine learning models or algorithms instead of the data itself [<xref ref-type="bibr" rid="ref41">41</xref>-<xref ref-type="bibr" rid="ref43">43</xref>]. Several research articles from our analysis proposed ways of automating the process. Fu et al [<xref ref-type="bibr" rid="ref3">3</xref>] present a straightforward way of measuring the agreement of AI-generated synthetic data against a gold standard dataset. Zhang et al [<xref ref-type="bibr" rid="ref19">19</xref>] suggest that while such straightforward analysis may be valuable, it is insufficient to measure true fairness, and they go on to propose a method of measuring bias via Jansen-Shannon divergence, which can be calculated for comparisons of real-world and synthetic data. The latter article also suggests a way of preventing synthetic data drift through condition regularization (ie, minimizing contrastive loss by regularizing the synthetic dataset against a real dataset distribution) and fuzzying (ie, adding controlled noise to broaden the dataset distribution before the AI training phase). To our knowledge, this is the most recently proposed technique for fairness assessment in the field. More research is needed to validate and augment the technique. Whether through Jansen-Shannon divergence or alternative methods, we recommend that all future data assessment projects measure and report model performance and fairness for sensitive groups.</p><p>Finally, Garcia-a-de-Leon-Chocano et al [<xref ref-type="bibr" rid="ref24">24</xref>] propose a way of calculating data robustness. The calculation draws on comparing tolerable versus nontolerable issues with DQ, which may be particularly important prior to using the information. We highly suggest that DQ assessments conduct a robustness calculation immediately before calculating data performance measures for downstream applications, which will allow for timely intervention in the case of significant issues with data completeness, conformity, or plausibility that merit additional data collection, review, or imputation steps as discussed above. The above findings and recommendations are summarized in <xref ref-type="table" rid="table4">Table 4</xref>.</p><table-wrap id="t4" position="float"><label>Table 4.</label><caption><p>Recommendations for future EHR<sup><xref ref-type="table-fn" rid="table4fn1">a</xref></sup> data quality and performance assessments.</p></caption><table id="table4" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Issue</td><td align="left" valign="bottom">Recommendation</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="2"><bold>Article characteristics</bold></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Paucity of specialty-focused EHR data assessments</td><td align="left" valign="top">Incentivize (eg, through quality improvement initiatives and grants) more EHR data assessments, particularly in psychiatry, emergency medicine, and surgical specialties</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Incomplete reporting</td><td align="left" valign="top">Use standardized frameworks for measuring and reporting data quality and performance assessments (eg, <xref ref-type="table" rid="table1">Table 1</xref>)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Poor replicability</td><td align="left" valign="top">Describe DQA<sup><xref ref-type="table-fn" rid="table4fn2">b</xref></sup> methods in enough details such that they could be replicated by a research team at a different institution</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Limited generalizability</td><td align="left" valign="top">Use already available data quality tools and standards (eg, DQA Guidelines proposed by Weiskopf et al [<xref ref-type="bibr" rid="ref21">21</xref>]) before developing proprietary methodologies</td></tr><tr><td align="left" valign="top" colspan="2"><bold>DQA</bold></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Inconsistent methodologies</td><td align="left" valign="top">Analyze completeness, conformance, and plausibility at every DQA (completeness only may be applicable for quick data quality checks)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Data missingness and nonconformity</td><td align="left" valign="top">Use available AI-based data extraction algorithms (eg, Lee et al [<xref ref-type="bibr" rid="ref22">22</xref>]), and augment data using external and synthetic datasets (eg, Zhang et al [<xref ref-type="bibr" rid="ref19">19</xref>])</td></tr><tr><td align="left" valign="top" colspan="2"><bold>Data performance assessment</bold></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Inconsistent methodologies</td><td align="left" valign="top">Augment correctness or accuracy measurement with recency, fairness, stability, and shareability performance metrics</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>EHR data bias</td><td align="left" valign="top">Automate data fairness assessments by measuring agreement of AI-extracted data against an gold standard dataset (eg, manually extracted data) and preventing drift via condition fuzzying and regularization (eg, Zhang et al [<xref ref-type="bibr" rid="ref19">19</xref>])</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Timeliness of analysis</td><td align="left" valign="top">Calculate dataset robustness prior to detailed data quality and performance analysis (eg, as described by Garc&#x00ED;a-de-Le&#x00F3;n-Chocano et al [<xref ref-type="bibr" rid="ref24">24</xref>])</td></tr></tbody></table><table-wrap-foot><fn id="table4fn1"><p><sup>a</sup>EHR: electronic health record.</p></fn><fn id="table4fn2"><p><sup>b</sup>DQA: data quality assessment.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s4-5"><title>Further Recommendations</title><p>Based on the review and our team&#x2019;s experience with DQ improvement initiatives, we recommend that administrators minimize modifiable DQ and performance issues arising during extraction by first using Internet of Things devices (eg, &#x201C;smart&#x201D; patient beds and infusion pumps) that directly upload measurements or settings to the EHR instead of requiring manual data entry. Second, the EHR&#x2019;s interface should be anchored to a predefined data workflow and ontological structure agreed upon in collaboration with clinical and data administrators (eg, encounters start at the time of patient check-in instead of when a physician first sees the patient, and all encounter times are recorded in 1 location using standard units). Finally, the plausibility of automatically entered data should be periodically validated such that corrections can be made when necessary (eg, a minute-by-minute electrocardiogram plausibility check that can detect if an electrocardiography lead falls off a patient&#x2019;s chest and needs to be replaced to record accurate measurements). Wherever possible, a reference data format (eg, electrocardiogram voltage between 0.5 and 5 mV) for the validation should be provided.</p><p>To minimize modifiable issues arising during data mappin<underline>g</underline>, we furthermore recommend first establishing rules for how to treat (1) &#x201C;missing,&#x201D; (2) &#x201C;modified,&#x201D; or (3) &#x201C;overlapping&#x201D; data, such as whether (1) fields with no value should be regarded as data points or artifacts; (2) data points that have been subsequently modified should be updated or retained; and (3) one data source should take precedence over another in case of duplicate records (eg, weight recordings measured by weighing scale should supersede those measured by a hospital bed). Finally, standards for parent-child encounters should be instituted (eg, if a postoperative outpatient clinic visit should be assigned as a unique encounter or as a child encounter of the parent surgery visit).</p><p>The provenance of outside facility records, which can be used to identify potential issues with externally collected data, should also be maintained (eg, keeping records of where and when outside laboratory measures were taken in order to identify potential issues with more or less accurate laboratory techniques).</p></sec><sec id="s4-6"><title>Limitations</title><p>While this scoping review provides valuable insight into the existing literature on EHR DQ analytics, it has several limitations. Foremost, it is important to acknowledge the limited sample size of 154 articles using our original search criteria, and consequently also the limited number of 26 original research articles which were included in our final analysis after full-text review. Among these articles, there was significant heterogeneity in settings and outcomes of interest, which may limit the validity of direct comparisons between the studies, as well as the generalizability of our findings. The review was furthermore restricted to articles available in the PubMed library, which may introduce a potential publication bias, as well as to articles available only in English, which may introduce a language bias to our study selection and subsequent analysis. Finally, while the review focused on EHR DQ and performance assessments, it did not include adjacent areas that may have a pronounced impact on clinical data recording and use such as EHR implementation or use. Future research should consider broader inclusion criteria and explore additional dimensions of EHR DQ to provide a more comprehensive understanding of this important topic.</p></sec><sec id="s4-7"><title>Conclusions</title><p>The findings of this scoping review highlight the importance of EHR DQ analysis in ensuring the accuracy and reliability of clinical data. Our review identified a need for specialty-specific data assessment initiatives, particularly in the fields of psychiatry, emergency medicine, and surgery. We additionally identified a need for standardizing DQ reporting to enhance the replicability and generalizability of outcomes in the field. Based on our review of the existing literature, we recommend analyzing DQ in terms of completeness, conformance, and plausibility; data performance in terms of correctness; and use case&#x2013;specific metrics such as recency, fairness, stability, and shareability. Notably, our review demonstrated several examples of DQ improvement with the use of AI-enhanced data extraction and supplementation techniques. Future efforts in augmenting DQ through AI should make use of data fairness assessments to prevent the introduction of synthetic data bias.</p></sec></sec></body><back><ack><p>TOB was supported by the National Institutes of Health (NIH; OT2 OD032701); the National Institute of Diabetes and Digestive and Kidney Diseases (NIH/NIDDK; K01 DK120784 and R01 DK121730); the National Institute of General Medical Sciences (NIH/NIGMS; R01 GM110240 and R01 GM149657); the National Institute of Biomedical Imaging and Bioengineering (NIH/NIBIB; R01 EB029699); the National Institute of Neurological Disorders and Stroke (NIH/NINDS; R01 NS120924); University of Florida (UF) Research (DRPD-ROSF2023 [00132783]); and the University of Florida Clinical and Translational Science Institute (AWD10247), which was supported in part by the NIH National Center for Advancing Translational Sciences (UL1TR001427). AB was supported by the NIH (OT2 OD032701), the National Institute of General Medical Sciences (NIH/NIGMS; R01 GM110240 and R01 GM149657), the National Institute of Biomedical Imaging and Bioengineering (NIH/NIBIB; R01 EB029699), the National Institute of Neurological Disorders and Stroke (NIH/NINDS; R01 NS120924), and the National Institute of Diabetes and Digestive and Kidney Diseases (NIH/NIDDK; R01 DK121730). TJL was supported by the National Institute of General Medical Sciences of the National Institutes of Health (R01 GM149657). BS was supported by the NIH (OT2 OD032701), by the National Institute of Diabetes and Digestive and Kidney Diseases (NIH/NIDDK; R01 DK121730), and by the National Institute of General Medical Sciences (NIH/NIGMS; R01 GM110240 and R01 GM149657). JB was supported by the NIH (T32 GM008721). The funding sources had no role in the design and conduct of the study; collection, management, analysis, and interpretation of the data; preparation, review, or approval of the paper; and decision to submit the paper for publication. The content is solely the responsibility of the authors and does not necessarily represent the official views of the NIH and other funding sources.</p></ack><fn-group><fn fn-type="con"><p>YPP performed the investigation, data curation, and writing&#x2014;original draft, review, and editing. TRB contributed to investigation, data curation, and writing&#x2014;original draft. MMR performed data curation, investigation, and writing&#x2014;review and editing. ML performed investigation. RS contributed to investigation. ZG did the investigation, methodology, and writing&#x2014;review and editing. JB did the data curation, methodology, writing&#x2014;review and editing&#x2014;and supervision. TOB performed data curation, methodology, and supervision. BS performed data curation, methodology, and supervision. TJL contributed to data curation, methodology, and supervision. AB performed data curation, methodology, and supervision.</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">DPA</term><def><p>data performance assessment</p></def></def-item><def-item><term id="abb3">DQ</term><def><p>data quality</p></def></def-item><def-item><term id="abb4">DQA</term><def><p>data quality assessment</p></def></def-item><def-item><term id="abb5">EHR</term><def><p>electronic health record</p></def></def-item><def-item><term id="abb6">GAN</term><def><p>generative adversarial network</p></def></def-item><def-item><term id="abb7">MIRACUM</term><def><p>Medical Informatics in Research and Care in University Medicine</p></def></def-item><def-item><term id="abb8">PCORnet</term><def><p>The National Patient-Centered Clinical Research Network</p></def></def-item><def-item><term id="abb9">PRISMA-ScR</term><def><p>Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><collab>All of Us Research Program Investigators</collab><name name-style="western"><surname>Denny</surname><given-names>JC</given-names> </name><name name-style="western"><surname>Rutter</surname><given-names>JL</given-names> </name><etal/></person-group><article-title>The &#x201C;All of Us&#x201D; research program</article-title><source>N Engl J Med</source><year>2019</year><month>08</month><day>15</day><volume>381</volume><issue>7</issue><fpage>668</fpage><lpage>676</lpage><pub-id pub-id-type="doi">10.1056/NEJMsr1809937</pub-id><pub-id pub-id-type="medline">31412182</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Curtis</surname><given-names>MD</given-names> </name><name name-style="western"><surname>Griffith</surname><given-names>SD</given-names> </name><name name-style="western"><surname>Tucker</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Development and validation of a high-quality composite real-world mortality endpoint</article-title><source>Health Serv Res</source><year>2018</year><month>12</month><volume>53</volume><issue>6</issue><fpage>4460</fpage><lpage>4476</lpage><pub-id pub-id-type="doi">10.1111/1475-6773.12872</pub-id><pub-id pub-id-type="medline">29756355</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fu</surname><given-names>S</given-names> </name><name name-style="western"><surname>Wen</surname><given-names>A</given-names> </name><name name-style="western"><surname>Pagali</surname><given-names>S</given-names> </name><etal/></person-group><article-title>The implication of latent information quality to the reproducibility of secondary use of electronic health records</article-title><source>Stud Health Technol Inform</source><year>2022</year><month>06</month><day>6</day><volume>290</volume><fpage>173</fpage><lpage>177</lpage><pub-id pub-id-type="doi">10.3233/SHTI220055</pub-id><pub-id pub-id-type="medline">35672994</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Engel</surname><given-names>N</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>H</given-names> </name><name name-style="western"><surname>Jiang</surname><given-names>X</given-names> </name><etal/></person-group><article-title>EHR data quality assessment tools and issue reporting Workflows for the &#x201C;All of Us&#x201D; research program clinical data research network</article-title><source>AMIA Jt Summits Transl Sci Proc</source><year>2022</year><month>05</month><volume>2022</volume><fpage>186</fpage><lpage>195</lpage><pub-id pub-id-type="medline">35854725</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kapsner</surname><given-names>LA</given-names> </name><name name-style="western"><surname>Mang</surname><given-names>JM</given-names> </name><name name-style="western"><surname>Mate</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Linking a consortium-wide data quality assessment tool with the MIRACUM metadata repository</article-title><source>Appl Clin Inform</source><year>2021</year><month>08</month><volume>12</volume><issue>4</issue><fpage>826</fpage><lpage>835</lpage><pub-id pub-id-type="doi">10.1055/s-0041-1733847</pub-id><pub-id pub-id-type="medline">34433217</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mohamed</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Song</surname><given-names>X</given-names> </name><name name-style="western"><surname>McMahon</surname><given-names>TM</given-names> </name><etal/></person-group><article-title>Tailoring rule-based data quality assessment to the Patient-Centered Outcomes Research Network (PCORnet) Common Data Model (CDM)</article-title><source>AMIA Annu Symp Proc</source><year>2023</year><month>04</month><day>29</day><volume>2022</volume><fpage>775</fpage><lpage>784</lpage><pub-id pub-id-type="medline">37128433</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="web"><article-title>Becoming the trusted open science community built with standardised health data via a European federated network</article-title><source>European Health Data &#x0026; Evidence Network</source><access-date>2024-10-23</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.ehden.eu/">https://www.ehden.eu/</ext-link></comment></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Aerts</surname><given-names>H</given-names> </name><name name-style="western"><surname>Kalra</surname><given-names>D</given-names> </name><name name-style="western"><surname>S&#x00E1;ez</surname><given-names>C</given-names> </name><etal/></person-group><article-title>Quality of hospital electronic health record (EHR) data based on the International Consortium for Health Outcomes Measurement (ICHOM) in heart failure: pilot data quality assessment study</article-title><source>JMIR Med Inform</source><year>2021</year><month>08</month><day>4</day><volume>9</volume><issue>8</issue><fpage>e27842</fpage><pub-id pub-id-type="doi">10.2196/27842</pub-id><pub-id pub-id-type="medline">34346902</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Puttkammer</surname><given-names>N</given-names> </name><name name-style="western"><surname>Baseman</surname><given-names>JG</given-names> </name><name name-style="western"><surname>Devine</surname><given-names>EB</given-names> </name><etal/></person-group><article-title>An assessment of data quality in a multi-site electronic medical record system in Haiti</article-title><source>Int J Med Inform</source><year>2016</year><month>02</month><volume>86</volume><fpage>104</fpage><lpage>116</lpage><pub-id pub-id-type="doi">10.1016/j.ijmedinf.2015.11.003</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bian</surname><given-names>J</given-names> </name><name name-style="western"><surname>Lyu</surname><given-names>T</given-names> </name><name name-style="western"><surname>Loiacono</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Assessing the practice of data quality evaluation in a national clinical data research network through a systematic scoping review in the era of real-world data</article-title><source>J Am Med Inform Assoc</source><year>2020</year><month>12</month><day>9</day><volume>27</volume><issue>12</issue><fpage>1999</fpage><lpage>2010</lpage><pub-id pub-id-type="doi">10.1093/jamia/ocaa245</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chan</surname><given-names>KS</given-names> </name><name name-style="western"><surname>Fowles</surname><given-names>JB</given-names> </name><name name-style="western"><surname>Weiner</surname><given-names>JP</given-names> </name></person-group><article-title>Review: electronic health records and the reliability and validity of quality measures: a review of the literature</article-title><source>Med Care Res Rev</source><year>2010</year><month>10</month><volume>67</volume><issue>5</issue><fpage>503</fpage><lpage>527</lpage><pub-id pub-id-type="doi">10.1177/1077558709359007</pub-id><pub-id pub-id-type="medline">20150441</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kahn</surname><given-names>MG</given-names> </name><name name-style="western"><surname>Callahan</surname><given-names>TJ</given-names> </name><name name-style="western"><surname>Barnard</surname><given-names>J</given-names> </name><etal/></person-group><article-title>A harmonized data quality assessment terminology and framework for the secondary use of electronic health record data</article-title><source>EGEMS (Wash DC)</source><year>2016</year><volume>4</volume><issue>1</issue><fpage>1244</fpage><pub-id pub-id-type="doi">10.13063/2327-9214.1244</pub-id><pub-id pub-id-type="medline">27713905</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Weiskopf</surname><given-names>NG</given-names> </name><name name-style="western"><surname>Weng</surname><given-names>C</given-names> </name></person-group><article-title>Methods and dimensions of electronic health record data quality assessment: enabling reuse for clinical research</article-title><source>J Am Med Inform Assoc</source><year>2013</year><month>01</month><day>1</day><volume>20</volume><issue>1</issue><fpage>144</fpage><lpage>151</lpage><pub-id pub-id-type="doi">10.1136/amiajnl-2011-000681</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Penning</surname><given-names>M</given-names> </name><name name-style="western"><surname>Zozus</surname><given-names>M</given-names> </name></person-group><article-title>Analysis of anesthesia screens for rule-based data quality assessment opportunities</article-title><source>Stud Health Technol Inform</source><year>2019</year><volume>257</volume><fpage>473</fpage><lpage>478</lpage><pub-id pub-id-type="medline">30741242</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kelley</surname><given-names>TA</given-names> </name></person-group><article-title>International Consortium for Health Outcomes Measurement (ICHOM)</article-title><source>Trials</source><year>2015</year><month>12</month><volume>16</volume><issue>S3</issue><pub-id pub-id-type="doi">10.1186/1745-6215-16-S3-O4</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Malik</surname><given-names>P</given-names> </name><name name-style="western"><surname>Pathania</surname><given-names>M</given-names> </name><name name-style="western"><surname>Rathaur</surname><given-names>VK</given-names> </name></person-group><article-title>Overview of artificial intelligence in medicine</article-title><source>J Family Med Prim Care</source><year>2019</year><month>07</month><volume>8</volume><issue>7</issue><fpage>2328</fpage><lpage>2331</lpage><pub-id pub-id-type="doi">10.4103/jfmpc.jfmpc_440_19</pub-id><pub-id pub-id-type="medline">31463251</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rajpurkar</surname><given-names>P</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>E</given-names> </name><name name-style="western"><surname>Banerjee</surname><given-names>O</given-names> </name><name name-style="western"><surname>Topol</surname><given-names>EJ</given-names> </name></person-group><article-title>AI in health and medicine</article-title><source>Nat Med</source><year>2022</year><month>01</month><volume>28</volume><issue>1</issue><fpage>31</fpage><lpage>38</lpage><pub-id pub-id-type="doi">10.1038/s41591-021-01614-0</pub-id><pub-id pub-id-type="medline">35058619</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gardner</surname><given-names>A</given-names> </name><name name-style="western"><surname>Smith</surname><given-names>AL</given-names> </name><name name-style="western"><surname>Steventon</surname><given-names>A</given-names> </name><name name-style="western"><surname>Coughlan</surname><given-names>E</given-names> </name><name name-style="western"><surname>Oldfield</surname><given-names>M</given-names> </name></person-group><article-title>Ethical funding for trustworthy AI: proposals to address the responsibilities of funders to ensure that projects adhere to trustworthy AI practice</article-title><source>AI Ethics</source><year>2022</year><month>05</month><volume>2</volume><issue>2</issue><fpage>277</fpage><lpage>291</lpage><pub-id pub-id-type="doi">10.1007/s43681-021-00069-w</pub-id><pub-id pub-id-type="medline">34790951</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Yan</surname><given-names>C</given-names> </name><name name-style="western"><surname>Malin</surname><given-names>BA</given-names> </name></person-group><article-title>Keeping synthetic patients on track: feedback mechanisms to mitigate performance drift in longitudinal health data simulation</article-title><source>J Am Med Inform Assoc</source><year>2022</year><month>10</month><day>7</day><volume>29</volume><issue>11</issue><fpage>1890</fpage><lpage>1898</lpage><pub-id pub-id-type="doi">10.1093/jamia/ocac131</pub-id><pub-id pub-id-type="medline">35927974</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ozonze</surname><given-names>O</given-names> </name><name name-style="western"><surname>Scott</surname><given-names>PJ</given-names> </name><name name-style="western"><surname>Hopgood</surname><given-names>AA</given-names> </name></person-group><article-title>Automating electronic health record data quality assessment</article-title><source>J Med Syst</source><year>2023</year><month>02</month><day>13</day><volume>47</volume><issue>1</issue><fpage>23</fpage><pub-id pub-id-type="doi">10.1007/s10916-022-01892-2</pub-id><pub-id pub-id-type="medline">36781551</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Weiskopf</surname><given-names>NG</given-names> </name><name name-style="western"><surname>Bakken</surname><given-names>S</given-names> </name><name name-style="western"><surname>Hripcsak</surname><given-names>G</given-names> </name><name name-style="western"><surname>Weng</surname><given-names>C</given-names> </name></person-group><article-title>A data quality assessment guideline for electronic health record data reuse</article-title><source>EGEMS (Wash DC)</source><year>2017</year><month>09</month><day>4</day><volume>5</volume><issue>1</issue><fpage>14</fpage><pub-id pub-id-type="doi">10.5334/egems.218</pub-id><pub-id pub-id-type="medline">29881734</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lee</surname><given-names>RY</given-names> </name><name name-style="western"><surname>Kross</surname><given-names>EK</given-names> </name><name name-style="western"><surname>Torrence</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Assessment of natural language processing of electronic health records to measure goals-of-care discussions as a clinical trial outcome</article-title><source>JAMA Netw Open</source><year>2023</year><month>03</month><day>1</day><volume>6</volume><issue>3</issue><fpage>e231204</fpage><pub-id pub-id-type="doi">10.1001/jamanetworkopen.2023.1204</pub-id><pub-id pub-id-type="medline">36862411</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tricco</surname><given-names>AC</given-names> </name><name name-style="western"><surname>Lillie</surname><given-names>E</given-names> </name><name name-style="western"><surname>Zarin</surname><given-names>W</given-names> </name><etal/></person-group><article-title>PRISMA extension for Scoping Reviews (PRISMA-SCR): checklist and explanation</article-title><source>Ann Intern Med</source><year>2018</year><month>10</month><day>2</day><volume>169</volume><issue>7</issue><fpage>467</fpage><lpage>473</lpage><pub-id pub-id-type="doi">10.7326/M18-0850</pub-id><pub-id pub-id-type="medline">30178033</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Garc&#x00ED;a-de-Le&#x00F3;n-Chocano</surname><given-names>R</given-names> </name><name name-style="western"><surname>S&#x00E1;ez</surname><given-names>C</given-names> </name><name name-style="western"><surname>Mu&#x00F1;oz-Soler</surname><given-names>V</given-names> </name><name name-style="western"><surname>Oliver-Roig</surname><given-names>A</given-names> </name><name name-style="western"><surname>Garc&#x00ED;a-de-Le&#x00F3;n-Gonz&#x00E1;lez</surname><given-names>R</given-names> </name><name name-style="western"><surname>Garc&#x00ED;a-G&#x00F3;mez</surname><given-names>JM</given-names> </name></person-group><article-title>Robust estimation of infant feeding indicators by data quality assessment of longitudinal electronic health records from birth up to 18 months of life</article-title><source>Comput Methods Programs Biomed</source><year>2021</year><month>08</month><volume>207</volume><fpage>106147</fpage><pub-id pub-id-type="doi">10.1016/j.cmpb.2021.106147</pub-id><pub-id pub-id-type="medline">34020376</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sirgo</surname><given-names>G</given-names> </name><name name-style="western"><surname>Esteban</surname><given-names>F</given-names> </name><name name-style="western"><surname>G&#x00F3;mez</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Validation of the ICU-DaMa tool for automatically extracting variables for minimum dataset and quality indicators: the importance of data quality assessment</article-title><source>Int J Med Inform</source><year>2018</year><month>04</month><volume>112</volume><fpage>166</fpage><lpage>172</lpage><pub-id pub-id-type="doi">10.1016/j.ijmedinf.2018.02.007</pub-id><pub-id pub-id-type="medline">29500016</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Toftdahl</surname><given-names>AKS</given-names> </name><name name-style="western"><surname>Pape-Haugaard</surname><given-names>LB</given-names> </name><name name-style="western"><surname>Palsson</surname><given-names>TS</given-names> </name><name name-style="western"><surname>Villumsen</surname><given-names>M</given-names> </name></person-group><article-title>Collect once - use many times: the research potential of low back pain patients&#x2019; municipal electronic healthcare records</article-title><source>Stud Health Technol Inform</source><year>2018</year><volume>247</volume><fpage>211</fpage><lpage>215</lpage><pub-id pub-id-type="medline">29677953</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tian</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Han</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>P</given-names> </name><name name-style="western"><surname>An</surname><given-names>J</given-names> </name><name name-style="western"><surname>Lu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Duan</surname><given-names>H</given-names> </name></person-group><article-title>Application of openEHR archetypes to automate data quality rules for electronic health records: a case study</article-title><source>BMC Med Inform Decis Mak</source><year>2021</year><month>04</month><day>3</day><volume>21</volume><issue>1</issue><fpage>113</fpage><pub-id pub-id-type="doi">10.1186/s12911-021-01481-2</pub-id><pub-id pub-id-type="medline">33812388</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tian</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Han</surname><given-names>Z</given-names> </name><name name-style="western"><surname>An</surname><given-names>J</given-names> </name><name name-style="western"><surname>Lu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Duan</surname><given-names>H</given-names> </name></person-group><article-title>Representing rules for clinical data quality assessment based on openEHR guideline definition language</article-title><source>Stud Health Technol Inform</source><year>2019</year><month>08</month><day>21</day><volume>264</volume><fpage>1606</fpage><lpage>1607</lpage><pub-id pub-id-type="doi">10.3233/SHTI190557</pub-id><pub-id pub-id-type="medline">31438254</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Estiri</surname><given-names>H</given-names> </name><name name-style="western"><surname>Stephens</surname><given-names>KA</given-names> </name><name name-style="western"><surname>Klann</surname><given-names>JG</given-names> </name><name name-style="western"><surname>Murphy</surname><given-names>SN</given-names> </name></person-group><article-title>Exploring completeness in clinical data research networks with DQe-c</article-title><source>J Am Med Inform Assoc</source><year>2018</year><month>01</month><day>1</day><volume>25</volume><issue>1</issue><fpage>17</fpage><lpage>24</lpage><pub-id pub-id-type="doi">10.1093/jamia/ocx109</pub-id><pub-id pub-id-type="medline">29069394</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mang</surname><given-names>JM</given-names> </name><name name-style="western"><surname>Seuchter</surname><given-names>SA</given-names> </name><name name-style="western"><surname>Gulden</surname><given-names>C</given-names> </name><etal/></person-group><article-title>DQAgui: a graphical user interface for the MIRACUM data quality assessment tool</article-title><source>BMC Med Inform Decis Mak</source><year>2022</year><month>08</month><day>11</day><volume>22</volume><issue>1</issue><fpage>213</fpage><pub-id pub-id-type="doi">10.1186/s12911-022-01961-z</pub-id><pub-id pub-id-type="medline">35953813</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sengupta</surname><given-names>S</given-names> </name><name name-style="western"><surname>Bachman</surname><given-names>D</given-names> </name><name name-style="western"><surname>Laws</surname><given-names>R</given-names> </name><etal/></person-group><article-title>Data quality assessment and multi-organizational reporting: tools to enhance network knowledge</article-title><source>EGEMS (Wash DC)</source><year>2019</year><month>03</month><day>29</day><volume>7</volume><issue>1</issue><fpage>8</fpage><pub-id pub-id-type="doi">10.5334/egems.280</pub-id><pub-id pub-id-type="medline">30972357</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Johnson</surname><given-names>SG</given-names> </name><name name-style="western"><surname>Speedie</surname><given-names>S</given-names> </name><name name-style="western"><surname>Simon</surname><given-names>G</given-names> </name><name name-style="western"><surname>Kumar</surname><given-names>V</given-names> </name><name name-style="western"><surname>Westra</surname><given-names>BL</given-names> </name></person-group><article-title>Application of an ontology for characterizing data quality for a secondary use of EHR data</article-title><source>Appl Clin Inform</source><year>2016</year><month>02</month><volume>7</volume><issue>1</issue><fpage>69</fpage><lpage>88</lpage><pub-id pub-id-type="doi">10.4338/ACI-2015-08-RA-0107</pub-id><pub-id pub-id-type="medline">27081408</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lee</surname><given-names>K</given-names> </name><name name-style="western"><surname>Weiskopf</surname><given-names>N</given-names> </name><name name-style="western"><surname>Pathak</surname><given-names>J</given-names> </name></person-group><article-title>A framework for data quality assessment in clinical research datasets</article-title><source>AMIA Annu Symp Proc</source><year>2018</year><month>04</month><volume>2017</volume><fpage>1080</fpage><lpage>1089</lpage><pub-id pub-id-type="medline">29854176</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kiogou</surname><given-names>SD</given-names> </name><name name-style="western"><surname>Chi</surname><given-names>CL</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>R</given-names> </name><name name-style="western"><surname>Ma</surname><given-names>S</given-names> </name><name name-style="western"><surname>Adam</surname><given-names>TJ</given-names> </name></person-group><article-title>Clinical data cohort quality improvement: the case of the medication data in the University of Minnesota&#x2019;s clinical data repository</article-title><source>AMIA Jt Summits Transl Sci Proc</source><year>2022</year><month>05</month><day>23</day><volume>2022</volume><fpage>293</fpage><lpage>302</lpage><pub-id pub-id-type="medline">35854717</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Thuraisingam</surname><given-names>S</given-names> </name><name name-style="western"><surname>Chondros</surname><given-names>P</given-names> </name><name name-style="western"><surname>Dowsey</surname><given-names>MM</given-names> </name><etal/></person-group><article-title>Assessing the suitability of general practice electronic health records for clinical prediction model development: a data quality assessment</article-title><source>BMC Med Inform Decis Mak</source><year>2021</year><month>10</month><day>30</day><volume>21</volume><issue>1</issue><fpage>297</fpage><pub-id pub-id-type="doi">10.1186/s12911-021-01669-6</pub-id><pub-id pub-id-type="medline">34717599</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Terry</surname><given-names>AL</given-names> </name><name name-style="western"><surname>Stewart</surname><given-names>M</given-names> </name><name name-style="western"><surname>Cejic</surname><given-names>S</given-names> </name><etal/></person-group><article-title>A basic model for assessing primary health care electronic medical record data quality</article-title><source>BMC Med Inform Decis Mak</source><year>2019</year><month>02</month><day>12</day><volume>19</volume><issue>1</issue><fpage>30</fpage><pub-id pub-id-type="doi">10.1186/s12911-019-0740-0</pub-id><pub-id pub-id-type="medline">30755205</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Walker</surname><given-names>KL</given-names> </name><name name-style="western"><surname>Kirillova</surname><given-names>O</given-names> </name><name name-style="western"><surname>Gillespie</surname><given-names>SE</given-names> </name><etal/></person-group><article-title>Using the CER Hub to ensure data quality in a multi-institution smoking cessation study</article-title><source>J Am Med Inform Assoc</source><year>2014</year><volume>21</volume><issue>6</issue><fpage>1129</fpage><lpage>1135</lpage><pub-id pub-id-type="doi">10.1136/amiajnl-2013-002629</pub-id><pub-id pub-id-type="medline">24993545</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gadde</surname><given-names>MA</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Zozus</surname><given-names>M</given-names> </name><name name-style="western"><surname>Talburt</surname><given-names>JB</given-names> </name><name name-style="western"><surname>Greer</surname><given-names>ML</given-names> </name></person-group><article-title>Rules based data quality assessment on claims database</article-title><source>Stud Health Technol Inform</source><year>2020</year><month>06</month><day>26</day><volume>272</volume><fpage>350</fpage><lpage>353</lpage><pub-id pub-id-type="doi">10.3233/SHTI200567</pub-id><pub-id pub-id-type="medline">32604674</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>H</given-names> </name><name name-style="western"><surname>Belitskaya-Levy</surname><given-names>I</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>F</given-names> </name><etal/></person-group><article-title>A statistical quality assessment method for longitudinal observations in electronic health record data with an application to the VA million veteran program</article-title><source>BMC Med Inform Decis Mak</source><year>2021</year><month>10</month><day>20</day><volume>21</volume><issue>1</issue><fpage>289</fpage><pub-id pub-id-type="doi">10.1186/s12911-021-01643-2</pub-id><pub-id pub-id-type="medline">34670548</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lewis</surname><given-names>AE</given-names> </name><name name-style="western"><surname>Weiskopf</surname><given-names>N</given-names> </name><name name-style="western"><surname>Abrams</surname><given-names>ZB</given-names> </name><etal/></person-group><article-title>Electronic health record data quality assessment and tools: a systematic review</article-title><source>J Am Med Inform Assoc</source><year>2023</year><month>09</month><day>25</day><volume>30</volume><issue>10</issue><fpage>1730</fpage><lpage>1740</lpage><pub-id pub-id-type="doi">10.1093/jamia/ocad120</pub-id><pub-id pub-id-type="medline">37390812</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="web"><person-group person-group-type="author"><collab>IBM</collab></person-group><article-title>AI Fairness 360 (AIF360)</article-title><source>GitHub</source><year>2023</year><access-date>2023-09-21</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://github.com/Trusted-AI/AIF360">https://github.com/Trusted-AI/AIF360</ext-link></comment></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="web"><person-group person-group-type="author"><collab>LinkedIn</collab></person-group><article-title>The Linkedin Fairness Toolkit (LiFT)</article-title><source>GitHub</source><year>2023</year><access-date>2023-09-21</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://github.com/linkedin/LiFT">https://github.com/linkedin/LiFT</ext-link></comment></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="web"><person-group person-group-type="author"><collab>Microsoft</collab></person-group><article-title>Responsible AI Toolbox</article-title><source>GitHub</source><year>2023</year><access-date>2023-09-21</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://github.com/microsoft/responsible-ai-toolbox">https://github.com/microsoft/responsible-ai-toolbox</ext-link></comment></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Search terms.</p><media xlink:href="medinform_v12i1e58130_app1.docx" xlink:title="DOCX File, 11 KB"/></supplementary-material><supplementary-material id="app2"><label>Multimedia Appendix 2</label><p>Study characteristics.</p><media xlink:href="medinform_v12i1e58130_app2.xlsx" xlink:title="XLSX File, 12 KB"/></supplementary-material><supplementary-material id="app3"><label>Checklist 1</label><p>PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) checklist.</p><media xlink:href="medinform_v12i1e58130_app3.docx" xlink:title="DOCX File, 54 KB"/></supplementary-material></app-group></back></article>