<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Nursing</journal-id><journal-id journal-id-type="publisher-id">nursing</journal-id><journal-id journal-id-type="index">33</journal-id><journal-title>JMIR Nursing</journal-title><abbrev-journal-title>JMIR Nursing</abbrev-journal-title><issn pub-type="epub">2562-7600</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v7i1e62678</article-id><article-id pub-id-type="doi">10.2196/62678</article-id><article-categories><subj-group subj-group-type="heading"><subject>Viewpoint</subject></subj-group></article-categories><title-group><article-title>Advancing AI Data Ethics in Nursing: Future Directions for Nursing Practice, Research, and Education</article-title></title-group><contrib-group><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Ball Dunlap</surname><given-names>Patricia A</given-names></name><degrees>MS, MSN, RN, NI-BC, CPHIMS</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Michalowski</surname><given-names>Martin</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib></contrib-group><aff id="aff1"><institution>School of Nursing, University of Minnesota</institution>, <addr-line>5-140 Weaver-Densford Hall, 308 Harvard Street SE</addr-line><addr-line>Minneapolis</addr-line><addr-line>MN</addr-line>, <country>United States</country></aff><aff id="aff2"><institution>Center for Digital Health, Mayo Clinic</institution>, <addr-line>Rochester</addr-line><addr-line>MN</addr-line>, <country>United States</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Borycki</surname><given-names>Elizabeth</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Park</surname><given-names>Claire Su-Yeon</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>He</surname><given-names>Lingxiao</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Patricia A Ball Dunlap, MS, MSN, RN, NI-BC, CPHIMS, School of Nursing, University of Minnesota, 5-140 Weaver-Densford Hall, 308 Harvard Street SE, Minneapolis, MN, 55455, United States, 1 612-624-5959; <email>patricia@rninformatics.ai</email></corresp></author-notes><pub-date pub-type="collection"><year>2024</year></pub-date><pub-date pub-type="epub"><day>25</day><month>10</month><year>2024</year></pub-date><volume>7</volume><elocation-id>e62678</elocation-id><history><date date-type="received"><day>28</day><month>05</month><year>2024</year></date><date date-type="rev-recd"><day>08</day><month>09</month><year>2024</year></date><date date-type="accepted"><day>13</day><month>09</month><year>2024</year></date></history><copyright-statement>&#x00A9; Patricia A Ball Dunlap, Martin Michalowski. Originally published in JMIR Nursing (<ext-link ext-link-type="uri" xlink:href="https://nursing.jmir.org">https://nursing.jmir.org</ext-link>), 25.10.2024. </copyright-statement><copyright-year>2024</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Nursing, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://nursing.jmir.org/">https://nursing.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://nursing.jmir.org/2024/1/e62678"/><abstract><p>The ethics of artificial intelligence (AI) are increasingly recognized due to concerns such as algorithmic bias, opacity, trust issues, data security, and fairness. Specifically, machine learning algorithms, central to AI technologies, are essential in striving for ethically sound systems that mimic human intelligence. These technologies rely heavily on data, which often remain obscured within complex systems and must be prioritized for ethical collection, processing, and usage. The significance of data ethics in achieving responsible AI was first highlighted in the broader context of health care and subsequently in nursing. This viewpoint explores the principles of data ethics, drawing on relevant frameworks and strategies identified through a formal literature review. These principles apply to real-world and synthetic data in AI and machine-learning contexts. Additionally, the data-centric AI paradigm is briefly examined, emphasizing its focus on data quality and the ethical development of AI solutions that integrate human-centered domain expertise. The ethical considerations specific to nursing are addressed, including 4 recommendations for future directions in nursing practice, research, and education and 2 hypothetical nurse-focused ethical case studies. The primary objectives are to position nurses to actively participate in AI and data ethics, thereby contributing to creating high-quality and relevant data for machine learning applications.</p></abstract><kwd-group><kwd>artificial intelligence</kwd><kwd>AI data ethics</kwd><kwd>data-centric AI</kwd><kwd>nurses</kwd><kwd>nursing informatics</kwd><kwd>machine learning</kwd><kwd>data literacy</kwd><kwd>health care AI</kwd><kwd>responsible AI</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Artificial intelligence (AI) has become increasingly popular in the United States and globally. Major US media outlets frequently report on AI, covering topics from job displacement concerns to its diverse and innovative applications across various industries. In the health care sector, where there is a vast amount of electronic administrative and clinical data, the adoption and application of AI technology are expected to grow significantly, with projections suggesting a market increase to approximately US$208 million by 2030 and a compounded annual growth rate of 38.5% worldwide [<xref ref-type="bibr" rid="ref1">1</xref>]. Moreover, AI ethics has gained more recognition due to the negative outcomes and ethical issues related to algorithmic bias, lack of transparency, trust issues, data security, and fairness. A preliminary examination of a global AI ethics case registry indicates that AI incidents proliferate across multiple industry sectors [<xref ref-type="bibr" rid="ref2">2</xref>].</p><p>Machine learning algorithms, particularly those featured in headlines or academic papers about ethical violations, are often central to these discussions. For instance, large language models and other machine learning algorithms have been reported to generate incorrect medical results [<xref ref-type="bibr" rid="ref3">3</xref>-<xref ref-type="bibr" rid="ref5">5</xref>], exhibit bias [<xref ref-type="bibr" rid="ref5">5</xref>,<xref ref-type="bibr" rid="ref6">6</xref>], and fail to identify chronically ill Black American patients who require high-risk care management [<xref ref-type="bibr" rid="ref5">5</xref>,<xref ref-type="bibr" rid="ref7">7</xref>]. These reports effectively draw attention to the malfunctioning of AI technologies. However, deeper analysis reveals that the underlying root cause of these failures often lies in the data used by these algorithms. Data integrity is fundamental to AI technology, as it powers these influential systems [<xref ref-type="bibr" rid="ref8">8</xref>-<xref ref-type="bibr" rid="ref10">10</xref>].</p><p>Poor quality or unrepresentative data in developing AI technologies can lead to significant issues, such as generative AI algorithms producing incorrect responses (known as &#x201C;hallucinations&#x201D;) and the degradation of machine learning model performance when encountering new data. These problems can disrupt operations and damage the public reputation of organizations, particularly in high-stakes environments like health care, thereby endangering patient outcomes and safety. Thus, data must be prioritized and scrutinized in discussions about AI, emphasizing the importance of ethical data collection, processing, and usage. Data ethics are critical for developing well-designed AI solutions and achieving responsible AI in health care, especially in technologies tailored for nursing.</p><p>Progressing from the existing literature, this viewpoint paper explores the principles of data ethics and their critical role in achieving responsible AI in health care and nursing. It also presents existing ethical data frameworks and strategies, emphasizing how data operations and usage contribute to generating high-quality data sets. These data sets are essential for training AI technologies to meet their intended performance and value objectives. The objectives of our viewpoints are to (1) introduce the principles of data ethics and relevant frameworks within the health care context; (2) examine how data-centric AI methodologies uphold the principles of data ethics by fostering the creation of high-quality data sets for machine learning; and (3) discuss the importance of AI data ethics, data-centric AI, and data quality for nursing practice, education, and research, including offering recommendations for future directions. We advocate further exploration and discussion of AI data ethics among nurses and nurse informatics researchers. With an effective organizational data governance structure and committed data-centric culture, we hypothesize that data ethics will significantly enhance AI development quality, thereby facilitating ethical and responsible AI solutions in practice, education, and research.</p></sec><sec id="s2"><title>Responsible AI</title><p>Before discussing AI data ethics in depth, it is pertinent to introduce the concept of responsible AI, as its principles greatly influence those of AI data ethics. Responsible AI aims to ensure that AI systems are designed, developed, and deployed in ethical, fair, transparent, accountable, and beneficial ways to all intended users [<xref ref-type="bibr" rid="ref11">11</xref>]. It is a multidimensional approach aiming to establish standards and values that prevent security issues, biases, and discrimination [<xref ref-type="bibr" rid="ref12">12</xref>]. Five principles underpin the creation of responsible AI solutions and they are (1) accountability; (2) sustainability; (3) bias, fairness, and privacy; (4) transparency and explainability; and (5) robustness, security, and safety [<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref14">14</xref>]. These principles also inform the principles of AI data ethics.</p></sec><sec id="s3"><title>AI Data Ethics</title><sec id="s3-1"><title>About AI Data Ethics and Its Role in AI</title><p>Data ethics is a critical aspect of responsible AI focusing on the ethical aspects of data operations, including data collection, processing, and use. It encompasses a broad range of ethical issues related to data handling, whether the data are structured or unstructured, across various modalities [<xref ref-type="bibr" rid="ref15">15</xref>-<xref ref-type="bibr" rid="ref17">17</xref>]. Existing research indicates that poor data ethics can have significant consequences. For instance, 57% of consumers reported they would stop doing business with companies that breached their trust through irresponsible data practices [<xref ref-type="bibr" rid="ref15">15</xref>]. Moreover, using suboptimal data to train machine learning algorithms can result in harmful outcomes for patients, such as misdiagnosis [<xref ref-type="bibr" rid="ref18">18</xref>], misidentification [<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref19">19</xref>], or data privacy breaches, even if the data were deidentified before use in machine learning algorithms [<xref ref-type="bibr" rid="ref8">8</xref>].</p><p>In the context of this viewpoint report, it is crucial to analyze data independently from the machine learning algorithms that process it. This separation underscores the need for strategies, governance, a data-literate organizational culture, and ethical principles that ensure the creation and use of high-quality data for AI technologies. As Radziwill [<xref ref-type="bibr" rid="ref20">20</xref>] noted, human error is often the root cause of bad data. Data may also reflect historical human biases embedded during their production, consumption, curation, and management. Therefore, data ethics is essential for addressing issues and unintended consequences stemming from poorly managed data.</p><p>Furthermore, there is a growing recognition of the importance of distinguishing data ethics from AI ethics, allowing for ethical decisions, specifically in data handling and the AI algorithms that use this data [<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref21">21</xref>]. This approach emphasizes a data-centric perspective that focuses on the moral dimensions of data [<xref ref-type="bibr" rid="ref16">16</xref>] and the ethical principles governing real-world and synthetic data [<xref ref-type="bibr" rid="ref22">22</xref>].</p></sec><sec id="s3-2"><title>Ethical Concerns Around Real-World and Synthetic Data in AI</title><p>Generative and predictive machine learning algorithms require extensive data sets to achieve performance objectives. Data scarcity has emerged as a significant concern within the AI and research communities, primarily due to the depletion of real-world training and validation data sets necessary for AI development. Several factors contribute to this issue, such as website owners increasingly protecting their data by enforcing data consent requirements [<xref ref-type="bibr" rid="ref23">23</xref>]. It is projected that data scarcity could become a critical issue between 2026 and 2032 [<xref ref-type="bibr" rid="ref24">24</xref>]. Additionally, the use of real-world data for AI development encounters problems such as missing data, which demands either imputation or deletion.</p><p>The generation of synthetic data is being accelerated to address issues related to data scarcity, privacy, and consent [<xref ref-type="bibr" rid="ref22">22</xref>]. Synthetic data are expected to surpass real-world data by 2030 [<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref25">25</xref>]. In the health care sector, synthetic data are used for simulation and prediction research, health IT development, education, and training [<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref26">26</xref>]. Despite its benefits, generating synthetic data for AI development presents ethical dilemmas and risks, especially in high-stakes areas such as health care. For instance, although synthetic data can help represent diverse populations and reduce algorithmic biases, overreliance on such data can lead to challenges and unforeseen long-term effects of converting unrepresentative data into representative data [<xref ref-type="bibr" rid="ref22">22</xref>]. Thus, ethical concerns arise regarding the non-maleficence and fidelity of synthetic data&#x2014;whether it can address real-world disparities or prevent the dissemination of misinformation [<xref ref-type="bibr" rid="ref22">22</xref>].</p></sec><sec id="s3-3"><title>Principles of AI Data Ethics</title><p>Rhem [<xref ref-type="bibr" rid="ref8">8</xref>] identified eight principles of AI data ethics, which are summarized as (1) transparency: is there clarity regarding the use, purpose, storage, and protection of the collected data? (2) Fairness: does the data collection and usage avoid exacerbating existing inequalities or biases? (3) Privacy: does the data collection process respect individuals&#x2019; privacy and autonomy, potentially through informed consent? (4) Responsibility: are data collectors and users accountable for ethical data collection and usage, including any harm resulting from these processes? Are mitigation steps in place? (5) Security: are data stored and transmitted securely to prevent unauthorized access, use, or disclosure? (6) Inclusivity: does the data collection and usage process ensure the inclusion of diverse perspectives and experiences, especially those that are underrepresented? (7) Transparency in decision-making: are decisions based on explainable and interpretable data? (8) Continual assessment: does the organization continuously monitor and assess its data practices to ensure they align with ethical principles [<xref ref-type="bibr" rid="ref8">8</xref>]?</p><p>Shanley et al [<xref ref-type="bibr" rid="ref22">22</xref>] were inspired by the five principles of responsible AI to initiate discussions on data ethics and synthetic data in AI. They proposed five principles to govern synthetic data (1) responsibility; (2) non-maleficence; (3) privacy; (4) transparency; and (5) justice, fairness, and equity [<xref ref-type="bibr" rid="ref22">22</xref>]. These principles closely align with the globally recognized principles of AI&#x2014;responsibility, non-maleficence, privacy, transparency, and justice and fairness [<xref ref-type="bibr" rid="ref27">27</xref>]. Moreover, the ethical principles suggested by Shanley et al [<xref ref-type="bibr" rid="ref22">22</xref>] and their associated questions correspond with Rhem&#x2019;s [<xref ref-type="bibr" rid="ref8">8</xref>] AI data ethical principles. This alignment is detailed in <xref ref-type="table" rid="table1">Table 1</xref>, which includes an adaptation of Rhem&#x2019;s [<xref ref-type="bibr" rid="ref8">8</xref>] principle of fairness.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Summary of data ethical principles for AI<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup>.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Principles of AI data ethics</td><td align="left" valign="bottom">Rhem [<xref ref-type="bibr" rid="ref8">8</xref>] questions</td><td align="left" valign="bottom">Shanley et al [<xref ref-type="bibr" rid="ref22">22</xref>] questions</td><td align="left" valign="bottom">New addition</td></tr></thead><tbody><tr><td align="left" valign="top">Transparency</td><td align="left" valign="top">Transparency: is there clarity regarding the use, purpose, storage, and protection of the collected data?</td><td align="left" valign="top">Transparency: how well does the synthetic data capture the phenomena it supposedly represents? How does the synthetic data deviate from the &#x201C;real&#x201D; data? What were the considerations when mitigating biases and how were they mitigated?</td><td align="left" valign="top">N/A<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup></td></tr><tr><td align="left" valign="top">Just, fair, and equitable data operations (DataOps)</td><td align="left" valign="top">Fairness: does the data collection and usage avoid exacerbating existing inequalities or biases?</td><td align="left" valign="top">Justice, fairness, and equity: are the underrepresented group&#x2019;s diversity and emerging novelties adequately considered? How are the developers held accountable for watching new characteristics, traits, or phenomena emerging within the synthetic data set? What is the process of alerting developers of overreliance on synthetic data for groups or populations where data collection is more challenging or costly?</td><td align="left" valign="top">N/A</td></tr><tr><td align="left" valign="top">Privacy</td><td align="left" valign="top">Privacy: does the data collection process respect individuals&#x2019; privacy and autonomy, potentially through informed consent?</td><td align="left" valign="top">Privacy: what data privacy policies need to be used for synthetic data set generation and use, including who is responsible for the policies? How do we obtain meaningful consent from the individuals and communities impacted by using their data for synthetic data set generation? What notions of data ownership should pertain to synthetic data set creation?</td><td align="left" valign="top">N/A</td></tr><tr><td align="left" valign="top">Responsibility</td><td align="left" valign="top">Responsibility: are data collectors and users accountable for ethical data collection and usage, including any harm resulting from these processes? Are mitigation steps in place?</td><td align="left" valign="top">Responsibility: who decides when and for what purpose synthetic data set generation is justified? When are real-world data necessary, and when is it appropriate to partially apply synthetic data sets?<break/>Suppose synthetic data entails accounting for additional considerations during the decision-making process. Does its use imply new or different responsibilities for those involved in the AI supply value chain? What does this mean for the roles, responsibilities, and decision-making processes of those involved in generating and using the synthetic data?</td><td align="left" valign="top">N/A</td></tr><tr><td align="left" valign="top">Security</td><td align="left" valign="top">Security: are data stored and transmitted securely to prevent unauthorized access, use, or disclosure?</td><td align="left" valign="top">N/A</td><td align="left" valign="top">N/A</td></tr><tr><td align="left" valign="top">Inclusivity</td><td align="left" valign="top">Inclusivity: does the data collection and usage process ensure the inclusion of diverse perspectives and experiences, especially those that are underrepresented?</td><td align="left" valign="top">Non-maleficence: what is the gap between the real world in which the AI is intended to operate and the synthetic world in which it was trained? What means and measures can we use to describe the gap adequately? And what vocabulary can we use to make sense of the uncollected data in the real world regarding its status vis-&#x00E0;-vis knowledge or truth claims? What is the potential for intentional or unintentional misuse of synthetic data?</td><td align="left" valign="top">N/A</td></tr><tr><td align="left" valign="top">Transparency in decision-making</td><td align="left" valign="top">Transparency in decision-making: are decisions based on explainable and interpretable data?</td><td align="left" valign="top">Transparency: see the transparency principle that is provided at the beginning of the table.</td><td align="left" valign="top">N/A</td></tr><tr><td align="left" valign="top">Continual assessment</td><td align="left" valign="top">Continual assessment: does the organization continuously monitor and assess its data practices to ensure they align with ethical principles?</td><td align="left" valign="top">N/A</td><td align="left" valign="top">N/A</td></tr><tr><td align="left" valign="top">Safety</td><td align="left" valign="top">N/A</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Do data operation activities and processes consider, identify, and mitigate risks associated with preventing danger, risk, or injury to individuals (patients)?</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>AI: artificial intelligence.</p></fn><fn id="table1fn2"><p><sup>b</sup>N/A: not applicable.</p></fn></table-wrap-foot></table-wrap><p>Patient safety is paramount in health care and nursing, encompassing the protection of patients from events such as danger, risk, or injury. Security concerns freedom from danger or threats, whereas privacy analyzes the unauthorized access or use of patient data. We suggest incorporating a ninth principle, safety, into Rhem&#x2019;s [<xref ref-type="bibr" rid="ref8">8</xref>] 8 principles of AI data ethics, as outlined in <xref ref-type="table" rid="table1">Table 1</xref>. A pertinent question for this new principle is &#x201C;Do data operation activities and processes consider, identify, and mitigate risks associated with preventing danger, risk, or injury to individuals (patients)?&#x201D; There is a dearth of consensus in the scientific and industrial literature from AI and health informatics research communities regarding definitions and practices that would guide AI data ethics. These definitions and shared principles are crucial for guiding the implementation and assessment of data ethics in clinical practice. As noted by Panai [<xref ref-type="bibr" rid="ref21">21</xref>], data ethics represents an underdeveloped area within organizations and lacks a clear definition. Similarly, AI data ethics is a latent or underexplored area in health care and nursing informatics scientific research.</p></sec><sec id="s3-4"><title>Ethical Data Frameworks</title><p>Numerous proposals exist for ethical data frameworks that support these principles. Floridi and Taddeo [<xref ref-type="bibr" rid="ref16">16</xref>] advocate for a macroethical data ethics framework, which aims to avoid narrow, ad-hoc approaches and enables organizations to develop solutions that optimize the societal benefits of data science. Furthermore, Marcovitch and Rancourt [<xref ref-type="bibr" rid="ref17">17</xref>] endorse standardized tools that facilitate the integration of data ethics accountability mechanisms, such as disclosure and transparency processes, at the organizational level. This is particularly important given the variations in legal frameworks across countries. Their proposed tools include the integration of data ethics into organizational culture, the establishment of data processes or management systems, a data governance structure, organizational transparency in ethical decision-making regarding the data supply chain, and a consistent method for demonstrating and verifying ethical data practices [<xref ref-type="bibr" rid="ref17">17</xref>].</p><p>Note that the frameworks proposed by Floridi and Taddeo [<xref ref-type="bibr" rid="ref16">16</xref>] and Marcovitch and Rancourt [<xref ref-type="bibr" rid="ref17">17</xref>] are not specifically tailored to the context of health care or nursing. The absence of a verifiable theoretical or conceptual AI data ethics framework that supports research, alongside a practical evaluation framework that translates effectively into practice, represents a significant gap in health and nursing informatics research. The ideas presented by Floridi and Taddeo [<xref ref-type="bibr" rid="ref16">16</xref>] and Marcovitch and Rancourt [<xref ref-type="bibr" rid="ref17">17</xref>] could serve as a foundation for developing such frameworks within the nursing field.</p></sec><sec id="s3-5"><title>Ethical Data Frameworks Challenges</title><p>Without a shared organizational data vision, strategy, and policies, implementing aspects of the proposed data ethical frameworks may be challenging, including risking data integrity. There needs to be clarity and knowledge about the responsibilities and liabilities of the people in charge of the data processes. The accountability and culture change should begin with the organization&#x2019;s executive team. Then, the executive team members intelligibly communicate the data cultural expectations, and policy changes to their departments, units, and teams. There is clarity about the responsibilities and liabilities of people who produce and consume the organizational data assets.</p><p>Nurses are producers and consumers of organizational data assets. For instance, they produce electronic health record (EHR) data and use them for nursing quality improvement initiatives. Positive deviance in effective data practices could be the impetus for remarkable cultural changes when poor data management practices are in place due to the absence of an organizational-level data vision, strategy, and policies. Suppose a nurse leader is passionate about the societal benefits of data science and AI, including taking the initiative to understand the importance of data quality. This individual becomes a change agent for their team. Their team&#x2019;s culture becomes data-centric, with patient safety and outcomes at the forefront. The team develops policies and standardized procedures that facilitate improved EHR data entry processes, which are less burdensome for the nurse but help them be accountable to the principles of AI data ethics. The team expresses knowledge about how data quality impacts patient care and the technologies they use in clinical settings. They observe technology as a mechanism to provide quality nursing care. Their attitudes and behaviors result in improved data quality for the nursing unit. This team is now a data vanguard. Other departments and nursing teams notice that this nursing unit outputs high-quality data, leading to improved insights for the unit. This &#x201C;positive deviant&#x201D; team&#x2019;s data practices led to informative unit-level reports, fewer data-related errors, and enhanced patient outcomes. Other nursing teams are curious and want to model this team&#x2019;s effective data procedures and practices. This influence could motivate the organization&#x2019;s nursing departments to follow suit.</p></sec></sec><sec id="s4"><title>Data-Centric AI</title><sec id="s4-1"><title>Model-Centric and Data-Centric AI Paradigms</title><p>AI technologies that leverage machine learning require substantial data for effective functioning. Machine learning is a subset of AI that enables computers to learn and adapt autonomously through algorithms and statistical models, with minimal or no human intervention. These systems demand extensive data volumes, with generative AI requiring even larger data sets and greater computing power to discern underlying patterns in the data. Historically, the development of prevalent AI technologies has adopted a model-centric strategy, prioritizing the machine learning algorithm or model as the primary focus for enhancing performance [<xref ref-type="bibr" rid="ref28">28</xref>]. In this approach, data-related activities such as curation, collection, and labeling are often deprioritized and occur only once, leading to potential ethical issues such as algorithmic bias and mispredictions.</p><p>The model-centric approach does not adequately address the principles of data ethics, as it overlooks the complexity, nuances, challenges, and accuracy of data, which are vital for improving the behavior of machine learning models [<xref ref-type="bibr" rid="ref29">29</xref>]. Recognizing these limitations, the AI industry is shifting toward a data-centric strategy, which places data&#x2014;the &#x201C;fuel&#x201D; of AI&#x2014;at the core of the development process [<xref ref-type="bibr" rid="ref28">28</xref>-<xref ref-type="bibr" rid="ref30">30</xref>]. This strategy emphasizes the importance of data quality to achieve high-performance machine learning models [<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref29">29</xref>]. Unlike the model-centric approach, data activities in the data-centric strategy are iterative, while model optimization remains static [<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref29">29</xref>]. This approach also promotes the involvement of domain experts to secure relevant, high-quality data sets for machine learning [<xref ref-type="bibr" rid="ref30">30</xref>].</p></sec><sec id="s4-2"><title>Data-Centric AI Role in Achieving Data Ethics and Responsible AI</title><p>The data-centric AI methodology involves developing, iterating, tracking, and maintaining the quality and integrity of AI systems&#x2019; data. It focuses on creating adequate training data, designing appropriate inference data, and ensuring data sustainability (establishing data lineage) [<xref ref-type="bibr" rid="ref29">29</xref>]. Given the importance of high-quality data, the data-centric approach aligns with the principles of data ethics and responsible AI. For instance, meticulous data curation and collection support ethical principles of fairness, responsibility, and transparency. The involvement of domain experts such as clinicians, clinical informaticists, and regulatory and privacy specialists promotes inclusivity, transparent decision-making, and privacy protection. These diverse perspectives help safeguard patient data privacy and prevent harm from substandard data. Considering data sets as a distinct and valuable product separate from the machine learning model also encourages health care organizations to establish continuous assessment protocols for their data collection and usage practices.</p></sec></sec><sec id="s5"><title>Significance of Data-Centric AI and AI Data Ethics in Nursing</title><p>In the broader health care context, we integrated the principles of data ethics and related data-centric AI strategies into nursing practices, emphasizing their role in creating high-quality data sets for machine learning. We now explore the significance of these concepts in nursing. First, AI has become increasingly prevalent in both nursing research and practice. A recent scoping review highlighted various AI activities and applications within nursing [<xref ref-type="bibr" rid="ref31">31</xref>]. Nurse researchers and informaticists need to understand the distinctions between model-centric and data-centric AI [<xref ref-type="bibr" rid="ref32">32</xref>] and their impacts on developing safe and effective AI technologies in health care settings, which influence care processes and workflows. This paper aims to establish a consensus on AI data ethics in nursing and motivate nurse informatics researchers to investigate further and discuss this crucial topic.</p><p>Second, numerous machine learning algorithms in health care use data from EHRs. Nurses, who are primary users of EHRs [<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref34">34</xref>], generate various electronic clinical documents detailing different aspects of patient care and progress (eg, admission assessments, nursing care plans, nursing education, and medication administration). We hypothesize that most frontline nurses are unaware of how their EHR data influences the performance and use of AI technologies. There is an existing awareness of the negative consequences of poorly designed EHR user interfaces on data quality in research and practice [<xref ref-type="bibr" rid="ref35">35</xref>-<xref ref-type="bibr" rid="ref38">38</xref>]. To effectively use EHR data in developing AI-driven clinical decision support systems, data collection and processing by the EHR user community must be optimized [<xref ref-type="bibr" rid="ref35">35</xref>]. Integrating AI data ethics into both practice and academic nursing curricula could enhance awareness about how EHR data are used by downstream systems and the role nurses play in generating high-quality EHR data for AI technologies. This approach introduces new research opportunities, such as examining the relationships between nurses&#x2019; data literacy and AI-related data quality.</p><p>Third, nurses who are more aware of data-centric AI and the principles of data ethics could increase their confidence in collaborating with data scientists, engineers, and other AI specialists on data-centric projects to produce meaningful, high-quality data and data sets. These efforts lead to development of high-performance machine learning models that align with nurses&#x2019; workflows. This applies to both real-world and synthetic data used in machine learning. Nurses provide valuable insights into patient safety and privacy and understand data collection at the point of care. Their contributions are crucial in helping health care organizations adhere to the principles of AI data ethics, develop responsible AI technologies that enhance their workflows, and support safe patient care.</p></sec><sec id="s6"><title>Nursing Practice and Research Implications</title><p>Machine learning algorithms, such as generative AI algorithms trained on poor-quality and unrepresentative data sets, can create significant political consequences in health care, such as exacerbating bias and health disparities. AI data ethics and data-centric AI represent emerging concepts in nursing. The nursing literature on these topics is limited, including discussions on the implications of using synthetic data to develop AI technologies tailored for nursing.</p><p>Finally, fictional nurse-oriented ethical case studies are presented in <xref ref-type="other" rid="box1">Textboxes 1</xref> and <xref ref-type="other" rid="box2">2</xref>. They illustrate potential ethical data breaches in real-world scenarios. These case studies can help nurses become informed about why AI data ethics should matter to them and examine their data practices, ensuring they perform their best in not becoming unknowingly enablers of data issues but high-data quality contributors and problem-solvers.</p><boxed-text id="box1"><title> Fictional nurse-specific case study #1: real-world data.</title><p><bold>Electronic health record (EHR) burnout leading to poor data entry with adverse data cascade effects</bold></p><p>Ava is a new graduate nurse in her fourth month of orienting on a busy adult medical-surgical unit. Six hours into her third consecutive 12-hour night shift, she must complete electronic clinical documentation for 4 patients in the EHR. Ava is sleep-deprived and highly stressed as she adjusts to becoming a competent med-surg nurse who can function without the supervision of a nurse preceptor. Additionally, her patient load was intense during this shift. She received a new admission from the telemetry floor about a few hours ago, a 70-year-old male patient named Carl. Carl arrived at Ava&#x2019;s unit moderately agitated and had a newly placed trach. Ava is ready to end her shift and get the much-needed respite. She must complete assessment documentation for Carl before the shift changes. The copy-forward feature was leveraged to accelerate Carl&#x2019;s assessment documentation, essentially copying some data elements previously documented by a telemetry nurse in the EHR. A data entry error embedded in the previous admission note is unknown to Ava. There is a mistake concerning Carl&#x2019;s medical history. In the telemetry admission note, the patient&#x2019;s smoking status was mistakenly documented as &#x201C;Non-smoker,&#x201D; although a relative communicated that Carl currently smokes. This incorrect value was pulled into Ava&#x2019;s shift assessment note via the copy-forward function. Moreover, due to exhaustion, Ava did not realize her assessment notes became bloated with irrelevant and duplicate data because of the copy-forward option. The shift assessment documentation copy-forward action occurred repeatedly by subsequent nurses caring for Carl during his hospitalization.</p><p>A few years later, a data scientist is requested to build a new machine learning model that will predict patients having a history of smoking and at risk for moderate to severe mental instability during their hospitalizations because of nicotine withdrawal and other socioeconomic factors. Furthermore, the final artificial intelligence (AI) solution will generate nursing care plans and patient education recommendations. This project was inspired by nurses&#x2019; desires to provide equitable care, taking precautionary steps to ensure these patients are comfortable and safe during their hospital stays. The erroneous data tied to Carl&#x2019;s previous hospital encounter were included in the training data set. An expert nurse was not engaged in the data collection and the data set validation processes.</p><p><bold>AI data ethical principle breach and brief commentary</bold>:</p><list list-type="bullet"><list-item><p>Responsibility: which data owners or consumers are accountable for the insufficient EHR data used to develop the machine learning model? What are the mitigation steps for mispredictions or generative AI hallucinations that may result in patient harm?</p></list-item><list-item><p>Inclusivity: how are we assured that the data sets represent the use case and intended patient population? This case study presents missing perspectives from nurse stakeholders, patient advocates, and critical decision makers.</p></list-item><list-item><p>Safety: a patient encounter with characteristics like Carl&#x2019;s data is predicted to be low risk, and hence, no recommendations were made by the AI solution. An expert nurse may catch the misprediction and take corrective steps. However, an inexperienced nurse may trust the prediction. This is a missed opportunity, and the proper treatment may be delayed or denied, potentially leading to an adverse patient outcome and reputational harm for the organization.</p></list-item></list></boxed-text><boxed-text id="box2"><title> Fictional nurse-specific case study #2: synthetic data.</title><p><bold>Generating synthetic data for health equity machine learning</bold></p><p>A data scientist is developing a nurse-specific machine learning model to predict whether patients are at risk of nonadherence to cardiac care at-home instructions, increasing their readmission risks. The discharge nurse provides and discusses these instructions with the patient before they are discharged from the hospital to home. The anonymized training and validation data sets have 500 and 300 observations, respectively, derived from the hospital&#x2019;s electronic health record (EHR) data. The data sets are not representative of a diverse patient population. This concern is significant to the data scientist because the data sets incorporate social drivers of health data elements (SDOH). After all, the project is funded by a federal grant with an initiative to improve health equity among disadvantaged patients in the United States. So, the data set must represent a diverse patient population. No formal organizational artificial intelligence (AI) data ethical policies or data ethics oversight committee exists to guide synthetic data creation. The data scientist does their best to add between 30 and 50 fictional observations, using the available training and validation data sets to guide the synthetic data modeling.</p><p><bold>AI data ethical principle breach and brief commentary</bold>:</p><list list-type="bullet"><list-item><p>Just, fair, and equitable data operations: the data scientist worked alone to create an assumable diverse data set. This decision should involve multiple key stakeholders and a vetting process to ensure the organization does not experience reputational harm and to protect patients from adverse outcomes. The potentially biased data set used to train the machine learning model may result in unforeseen algorithmic bias.</p></list-item><list-item><p>Responsibility: if an adverse patient outcome occurs because a nurse trusted the predictions made by this machine learning model, who is held responsible? What does the root cause analysis process look like?</p></list-item><list-item><p>Inclusivity: the &#x201C;diverse&#x201D; data set the data scientist developed may be insufficient and unrepresentative of the intended target population. What policies are guiding the decision around what is considered inclusive data? There are missing perspectives from key stakeholders.</p></list-item><list-item><p>Transparency in decision-making: can the data scientist adequately explain the decisions behind developing the data sets to nontechnical users? After making a prediction, does the AI solution provide interpretable results that could guide nurses about how it arrived at its conclusion?</p></list-item><list-item><p>Safety: has the data scientist documented the data activities, including the identified risks to nurses and their patients and the risks&#x2019; mitigation plans?</p></list-item></list></boxed-text></sec><sec id="s7"><title>Recommendations for Nursing</title><p>We present 4 recommendations to enable nurses to engage with and contribute to developing responsible AI technologies that align with their workflows and adhere to the principles of AI data ethics. Implementation of the recommendations could transform nursing care, practice, and education around data. They aim to prepare nurses for their future in practicing and learning in health care AI. Transformations include (1) AI technologies effectively reduce nurses&#x2019; burdensome documentation, (2) AI and data-literate nurses experiencing a reduction in technological fears like job displacement&#x2014;rather than fearing the technology, learn and use it to their advantage, (3) nurses advocating for their profession and patients by getting involved in the design of AI technologies, and (4) nurses&#x2019; improved data knowledge and management practices leading to insights that positively impact patient outcomes and the service they provide to their patients.</p><sec id="s7-1"><title>Recommendation 1: Data Ethics Engagement Necessitates Data Literacy in AI</title><p>As AI becomes increasingly integral to the daily responsibilities of nurses across various clinical and administrative settings, nurse leadership must promote a data-centric culture within the nursing profession. Leaders should serve as role models, emphasizing the importance of data quality in nursing practice. One approach to achieving this is by enhancing nurses&#x2019; understanding of AI and data. Nurses need to comprehend how downstream systems use the data they produce in EHR. It is crucial to reflect, correct, and evolve from the existing cultural norms and power dynamics that hinder effective EHR data collection, establishing a new culture that recognizes the importance of maintaining data quality at the point of care.</p><p>Additionally, nurse educators and researchers should strive to improve their AI and data literacy skills and develop new curricula to bridge the knowledge gap in AI and data among nursing students, particularly those enrolled in nurse informatics programs or those pursuing research in informatics. Data literacy is the ability to explore, read, write, understand, and communicate meaningfully within a specific context [<xref ref-type="bibr" rid="ref39">39</xref>,<xref ref-type="bibr" rid="ref40">40</xref>].</p><p>Improved data literacy facilitates discussions regarding data usage and ethics in AI. Establishing a common language that clearly defines data literacy and ethics in nursing is necessary, including the principles of AI data ethics to guide the development of nurse-specific AI technologies. Furthermore, a shared language and a solid foundation in data knowledge equip nurses to act as data stewards, engage in ethical discussions, innovate in data management, and collaborate with AI specialists to develop responsible, nurse-specific AI technologies. These initiatives align with the principles of inclusivity and responsibility in AI data ethics.</p></sec><sec id="s7-2"><title>Recommendation 2: Data-Oriented Culture Motivates Ethical Accountability Mechanisms</title><p>Creating and nurturing a data-oriented culture among nurses can enhance their involvement in ethical accountability mechanisms for AI data, as outlined by Marcovitch and Rancourt [<xref ref-type="bibr" rid="ref17">17</xref>]. However, the concept of data ethics remains poorly defined, leading to overlapping responsibilities and ambiguous accountabilities. This lack of clarity can hinder the effective detection of ethical violations in data use, as data ethics often merges into broader ethical principles that are insufficient at the granular level of data abstraction [<xref ref-type="bibr" rid="ref21">21</xref>]. Consequently, there is a need for specific job roles focused on how data ethics can inform ethical decisions regarding data and the machine learning algorithms that process it. These roles would involve establishing relevant policies and practices and advocating for the consumer&#x2014;here, the patient. Thus, the recommendation to establish a Chief Data Ethics Officer role emerges [<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref41">41</xref>]. The mandate of this officer is ethical rather than legal [<xref ref-type="bibr" rid="ref41">41</xref>], focusing on leading a support team responsible for drafting a code of data ethics, forming and managing a data ethics committee, and overseeing data-oriented ethical issues [<xref ref-type="bibr" rid="ref21">21</xref>].</p><p>With a data-centric organizational culture, data governance, and a Chief Data Ethics Officer, nurse leaders and researchers can explore and implement innovative, nursing-centric data roles. These roles are crucial for upholding the principles of AI data ethics in nursing practice and creating standardized data accountability tools, policies, and processes that effectively evaluate and measure AI data ethics in nurse-specific AI technologies. Job roles such as Chief Nurse Data Ethics Officer and Nurse Data Steward should be explored further.</p><p>Nurse educators should implement courses that prepare nurses for data-oriented roles in AI. Nursing degree programs at both undergraduate and graduate levels should be adapted to include relevant courses on AI, data ethics, foundational data science, and data literacy. Nurse scientists are encouraged to explore novel approaches, frameworks, and instruments that enable the integration of ethical accountability mechanisms and effectively assess the principles of AI data ethics reinforced by scientific evidence.</p></sec><sec id="s7-3"><title>Recommendation 3: Optimal Data Quality Is Conditioned on Domain Expertise Participation</title><p>Domain expertise is essential for generating high-quality data [<xref ref-type="bibr" rid="ref32">32</xref>], and human involvement is critical to successfully executing data-centric AI tasks [<xref ref-type="bibr" rid="ref29">29</xref>]. Nurses can use their knowledge of data-centric AI and data ethics to actively engage in all phases of AI development and positively leverage their expertise. Specifically, nurse informaticists have opportunities to participate in activities such as data annotation, labeling, and the verification and validation of data elements used in AI data sets. Furthermore, throughout the AI development lifecycle, nurses can play a pivotal role in identifying and addressing opaque data decisions that may affect frontline clinicians&#x2019; trust and usage of AI technologies, thereby enhancing transparency in decision-making processes. The active involvement of nurses in these data-centric AI activities, combined with their commitment to patient advocacy, supports the ethical principles of data privacy and promotes accountability and ownership of data collection, processing, and use.</p></sec><sec id="s7-4"><title>Recommendation 4: High-Quality, Ethical Data Curtails Health Care Political Consequences</title><p>Nurse leaders, educators, and researchers must recognize that implementing responsible AI technologies and initiatives in health care is complex and challenging, despite significant interest in AI ethics [<xref ref-type="bibr" rid="ref5">5</xref>]. This statement is not intended to deter nurses from exploring methods to develop ethical AI technologies that use high-quality, representative clinical data. Instead, it aims to raise awareness and encourage nurses to persevere through challenges, including maintaining patience and resilience during the change management process. AI technologies are often politically influenced, reflecting their designer&#x2019;s values, beliefs, and norms, as well as the data and data sets used to train them [<xref ref-type="bibr" rid="ref5">5</xref>,<xref ref-type="bibr" rid="ref42">42</xref>]. Political consequences may arise from data operations such as selection, labeling, preprocessing, and transformation [<xref ref-type="bibr" rid="ref42">42</xref>]. Nurses, working at various touchpoints across diverse settings in the health care ecosystem, bring unique perspectives to the AI discussion, which can help promote and advocate for the principles of AI data ethics.</p></sec><sec id="s7-5"><title>Conclusions</title><p>Working with data is time-consuming and challenging and often perceived as less exciting than developing machine learning models or AI technologies. Despite this, the foundational role of data in AI systems cannot be overstated; high-quality data are crucial for the performance and value of AI technologies in health care. The principles of AI data ethics aim to promote responsible AI and the creation of ethical AI technologies [<xref ref-type="bibr" rid="ref8">8</xref>]. The methods used to collect, store, use, and share data have profound implications for individuals, organizations, and society [<xref ref-type="bibr" rid="ref8">8</xref>].</p></sec></sec></body><back><fn-group><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">EHR</term><def><p>electronic health record</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="web"><article-title>AI in healthcare market size, share &#x0026; trends analysis report by component (hardware, services), by application, by end-use, by technology, by region, and segment forecasts, 2024 - 2030</article-title><source>Grand View Research</source><year>2023</year><access-date>2023-10-14</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.grandviewresearch.com/industry-analysis/artificial-intelligence-ai-healthcare-market">https://www.grandviewresearch.com/industry-analysis/artificial-intelligence-ai-healthcare-market</ext-link></comment></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="web"><article-title>AI ethics case studies &#x0026; AI incident registries</article-title><source>AI Ethicist</source><access-date>2023-12-28</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.aiethicist.org/ethics-cases-registries">https://www.aiethicist.org/ethics-cases-registries</ext-link></comment></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bhattacharyya</surname><given-names>M</given-names> </name><name name-style="western"><surname>Miller</surname><given-names>VM</given-names> </name><name name-style="western"><surname>Bhattacharyya</surname><given-names>D</given-names> </name><name name-style="western"><surname>Miller</surname><given-names>LE</given-names> </name></person-group><article-title>High rates of fabricated and inaccurate references in ChatGPT-generated medical content</article-title><source>Cureus</source><year>2023</year><month>05</month><volume>15</volume><issue>5</issue><fpage>e39238</fpage><pub-id pub-id-type="doi">10.7759/cureus.39238</pub-id><pub-id pub-id-type="medline">37337480</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Emsley</surname><given-names>R</given-names> </name></person-group><article-title>ChatGPT: these are not hallucinations - they&#x2019;re fabrications and falsifications</article-title><source>Schizophrenia (Heidelb)</source><year>2023</year><month>08</month><day>19</day><volume>9</volume><issue>1</issue><fpage>52</fpage><pub-id pub-id-type="doi">10.1038/s41537-023-00379-4</pub-id><pub-id pub-id-type="medline">37598184</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Siala</surname><given-names>H</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>Y</given-names> </name></person-group><article-title>SHIFTing artificial intelligence to be responsible in healthcare: a systematic review</article-title><source>Soc Sci Med</source><year>2022</year><month>03</month><volume>296</volume><fpage>114782</fpage><pub-id pub-id-type="doi">10.1016/j.socscimed.2022.114782</pub-id><pub-id pub-id-type="medline">35152047</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mittermaier</surname><given-names>M</given-names> </name><name name-style="western"><surname>Raza</surname><given-names>MM</given-names> </name><name name-style="western"><surname>Kvedar</surname><given-names>JC</given-names> </name></person-group><article-title>Bias in AI-based models for medical applications: challenges and mitigation strategies</article-title><source>NPJ Digit Med</source><year>2023</year><month>06</month><day>14</day><volume>6</volume><issue>1</issue><fpage>113</fpage><pub-id pub-id-type="doi">10.1038/s41746-023-00858-z</pub-id><pub-id pub-id-type="medline">37311802</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Obermeyer</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Powers</surname><given-names>B</given-names> </name><name name-style="western"><surname>Vogeli</surname><given-names>C</given-names> </name><name name-style="western"><surname>Mullainathan</surname><given-names>S</given-names> </name></person-group><article-title>Dissecting racial bias in an algorithm used to manage the health of populations</article-title><source>Science</source><year>2019</year><month>10</month><day>25</day><volume>366</volume><issue>6464</issue><fpage>447</fpage><lpage>453</lpage><pub-id pub-id-type="doi">10.1126/science.aax2342</pub-id><pub-id pub-id-type="medline">31649194</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Rhem</surname><given-names>AJ</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Miroslav</surname><given-names>R</given-names> </name></person-group><article-title>Ethical use of data in AI applications</article-title><source>Ethics - Scientific Research, Ethical Issues, Artificial Intelligence and Education</source><year>2023</year><edition>1</edition><publisher-name>IntechOpen</publisher-name><pub-id pub-id-type="doi">10.5772/intechopen.1001597</pub-id><pub-id pub-id-type="other">9781837695270</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Budach</surname><given-names>L</given-names> </name><name name-style="western"><surname>Feuerpfeil</surname><given-names>M</given-names> </name><name name-style="western"><surname>Ihde</surname><given-names>N</given-names> </name><etal/></person-group><article-title>The effects of data quality on machine learning performance</article-title><source>arXiv</source><comment>Preprint posted online on  Jul 29, 2022</comment><pub-id pub-id-type="doi">10.48550/arXiv.2207.14529</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Ng</surname><given-names>A</given-names> </name></person-group><article-title>A chat with Andrew on mlops: from model-centric to data-centric AI</article-title><source>YouTube</source><year>2021</year><access-date>2022-10-08</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.youtube.com/watch?v=06-AZXmwHjo">https://www.youtube.com/watch?v=06-AZXmwHjo</ext-link></comment></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>R</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>D</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>Z</given-names> </name></person-group><article-title>Editorial: responsible AI in healthcare: opportunities, challenges, and best practices</article-title><source>Front Comput Sci</source><year>2023</year><volume>5</volume><pub-id pub-id-type="doi">10.3389/fcomp.2023.1265902</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Merhi</surname><given-names>MI</given-names> </name></person-group><article-title>An assessment of the barriers impacting responsible artificial intelligence</article-title><source>Inf Syst Front</source><year>2023</year><month>06</month><volume>25</volume><issue>3</issue><fpage>1147</fpage><lpage>1160</lpage><pub-id pub-id-type="doi">10.1007/s10796-022-10276-3</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="web"><article-title>OECD AI principles overview</article-title><source>OECDAI Policy Observatory</source><access-date>2024-08-22</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://oecd.ai/en/ai-principles">https://oecd.ai/en/ai-principles</ext-link></comment></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="web"><article-title>Frequently asked questions: what is responsible artificial intelligence (rAI)?</article-title><source>Responsible Artificial Intelligence Institute</source><access-date>2024-08-22</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.responsible.ai/faq/">https://www.responsible.ai/faq/</ext-link></comment></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="web"><article-title>Data ethics: what is data ethics?</article-title><source>Cognizant</source><access-date>2023-11-28</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.cognizant.com/us/en/glossary/data-ethics">https://www.cognizant.com/us/en/glossary/data-ethics</ext-link></comment></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Floridi</surname><given-names>L</given-names> </name><name name-style="western"><surname>Taddeo</surname><given-names>M</given-names> </name></person-group><article-title>What is data ethics?</article-title><source>Philos Trans A Math Phys Eng Sci</source><year>2016</year><month>12</month><day>28</day><volume>374</volume><issue>2083</issue><fpage>20160360</fpage><pub-id pub-id-type="doi">10.1098/rsta.2016.0360</pub-id><pub-id pub-id-type="medline">28336805</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Marcovitch</surname><given-names>I</given-names> </name><name name-style="western"><surname>Rancourt</surname><given-names>E</given-names> </name></person-group><article-title>A data ethics framework for responsible responsive organizations in the digital world</article-title><source>Statistical J IAOS</source><year>2022</year><volume>38</volume><issue>4</issue><fpage>1161</fpage><lpage>1172</lpage><pub-id pub-id-type="doi">10.3233/SJI-220067</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Roberts</surname><given-names>M</given-names> </name><name name-style="western"><surname>Driggs</surname><given-names>D</given-names> </name><name name-style="western"><surname>Thorpe</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Common pitfalls and recommendations for using machine learning to detect and prognosticate for COVID-19 using chest radiographs and CT scans</article-title><source>Nat Mach Intell</source><year>2021</year><volume>3</volume><issue>3</issue><fpage>199</fpage><lpage>217</lpage><pub-id pub-id-type="doi">10.1038/s42256-021-00307-0</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Heaven</surname><given-names>WD</given-names> </name></person-group><article-title>Hundreds of AI tools have been built to catch covid. None of them helped</article-title><source>MIT Technology Review</source><year>2021</year><access-date>2023-12-02</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.technologyreview.com/2021/07/30/1030329/machine-learning-ai-failed-covid-hospital-diagnosis-pandemic/">https://www.technologyreview.com/2021/07/30/1030329/machine-learning-ai-failed-covid-hospital-diagnosis-pandemic/</ext-link></comment></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Radziwill</surname><given-names>NM</given-names> </name></person-group><source>Data, Strategy, Culture &#x0026; Power: Win with Data-Centric AI by Making Human Nature Work for You</source><year>2024</year><edition>1</edition><publisher-name>Lapis Lucera</publisher-name><pub-id pub-id-type="other">9780996916080</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Panai</surname><given-names>E</given-names> </name></person-group><article-title>The latent space of data ethics</article-title><source>AI &#x0026; Soc</source><year>2023</year><pub-id pub-id-type="doi">10.1007/s00146-023-01757-3</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shanley</surname><given-names>D</given-names> </name><name name-style="western"><surname>Hogenboom</surname><given-names>J</given-names> </name><name name-style="western"><surname>Lysen</surname><given-names>F</given-names> </name></person-group><article-title>Getting real about synthetic data ethics: are AI ethics principles a good starting point for synthetic data ethics?</article-title><source>EMBO Rep</source><year>2024</year><month>05</month><volume>25</volume><issue>5</issue><fpage>2152</fpage><lpage>2155</lpage><pub-id pub-id-type="doi">10.1038/s44319-024-00101-0</pub-id><pub-id pub-id-type="medline">38388694</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Longpre</surname><given-names>S</given-names> </name><name name-style="western"><surname>Mahari</surname><given-names>R</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Consent in crisis: the rapid decline of the AI data commons</article-title><source>Data Provenance Initiative</source><access-date>2024-08-22</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.dataprovenance.org/consent-in-crisis-paper">https://www.dataprovenance.org/consent-in-crisis-paper</ext-link></comment></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Villalobos</surname><given-names>P</given-names> </name><name name-style="western"><surname>Sevilla</surname><given-names>J</given-names> </name><name name-style="western"><surname>Heim</surname><given-names>L</given-names> </name><name name-style="western"><surname>Besiroglu</surname><given-names>T</given-names> </name><name name-style="western"><surname>Hobbhahn</surname><given-names>M</given-names> </name><name name-style="western"><surname>Ho</surname><given-names>A</given-names> </name></person-group><article-title>Will we run out of data? An analysis of the limits of scaling datasets in machine learning</article-title><source>arXiv</source><comment>Preprint posted online on  Oct 26, 2022</comment><pub-id pub-id-type="doi">10.48550/arXiv.2211.04325</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Linden</surname><given-names>A</given-names> </name></person-group><article-title>Is synthetic data the future of AI</article-title><source>Gartner</source><access-date>2024-08-22</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.gartner.com/en/newsroom/press-releases/2022-06-22-is-synthetic-data-the-future-of-ai">https://www.gartner.com/en/newsroom/press-releases/2022-06-22-is-synthetic-data-the-future-of-ai</ext-link></comment></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gonzales</surname><given-names>A</given-names> </name><name name-style="western"><surname>Guruswamy</surname><given-names>G</given-names> </name><name name-style="western"><surname>Smith</surname><given-names>SR</given-names> </name></person-group><article-title>Synthetic data in health care: a narrative review</article-title><source>PLOS Dig Health</source><year>2023</year><month>01</month><volume>2</volume><issue>1</issue><fpage>e0000082</fpage><pub-id pub-id-type="doi">10.1371/journal.pdig.0000082</pub-id><pub-id pub-id-type="medline">36812604</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jobin</surname><given-names>A</given-names> </name><name name-style="western"><surname>Ienca</surname><given-names>M</given-names> </name><name name-style="western"><surname>Vayena</surname><given-names>E</given-names> </name></person-group><article-title>The global landscape of AI ethics guidelines</article-title><source>Nat Mach Intell</source><year>2019</year><volume>1</volume><issue>9</issue><fpage>389</fpage><lpage>399</lpage><pub-id pub-id-type="doi">10.1038/s42256-019-0088-2</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Hamid</surname><given-names>OH</given-names> </name></person-group><article-title>From model-centric to data-centric AI: a paradigm shift or rather a complementary approach?</article-title><year>2022</year><conf-name>2022 8th International Conference on Information Technology Trends (ITT)</conf-name><conf-date>May 25-26, 2022</conf-date><conf-loc>Dubai, United Arab Emirates</conf-loc><pub-id pub-id-type="doi">10.1109/ITT56123.2022.9863935</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Zha</surname><given-names>D</given-names> </name><name name-style="western"><surname>Bhat</surname><given-names>ZP</given-names> </name><name name-style="western"><surname>Lai</surname><given-names>KH</given-names> </name><etal/></person-group><article-title>Data-centric artificial intelligence: a survey</article-title><source>arXiv</source><comment>Preprint posted online on  Mar 17, 2023</comment><pub-id pub-id-type="doi">10.48550/arXiv.2303.10158</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="web"><article-title>Data-centric AI vs. model-centric AI - everything you need know</article-title><source>Artificial Intelligence Board of America</source><year>2022</year><access-date>2022-12-14</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.artiba.org/blog/data-centric-ai-vs-model-centric-ai-everything-you-need-know">https://www.artiba.org/blog/data-centric-ai-vs-model-centric-ai-everything-you-need-know</ext-link></comment></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hwang</surname><given-names>GJ</given-names> </name><name name-style="western"><surname>Chang</surname><given-names>PY</given-names> </name><name name-style="western"><surname>Tseng</surname><given-names>WY</given-names> </name><name name-style="western"><surname>Chou</surname><given-names>CA</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>CH</given-names> </name><name name-style="western"><surname>Tu</surname><given-names>YF</given-names> </name></person-group><article-title>Research trends in artificial intelligence-associated nursing activities based on a review of academic studies published from 2001 to 2020</article-title><source>Comput Inform Nurs</source><year>2022</year><month>12</month><day>1</day><volume>40</volume><issue>12</issue><fpage>814</fpage><lpage>824</lpage><pub-id pub-id-type="doi">10.1097/CIN.0000000000000897</pub-id><pub-id pub-id-type="medline">36516032</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ball Dunlap</surname><given-names>PA</given-names> </name><name name-style="western"><surname>Nahm</surname><given-names>ES</given-names> </name><name name-style="western"><surname>Umberfield</surname><given-names>EE</given-names> </name></person-group><article-title>Data-centric machine learning in nursing: a concept clarification</article-title><source>Comput Inform Nurs</source><year>2024</year><month>05</month><day>1</day><volume>42</volume><issue>5</issue><fpage>325</fpage><lpage>333</lpage><pub-id pub-id-type="doi">10.1097/CIN.0000000000001102</pub-id><pub-id pub-id-type="medline">38241753</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Collins</surname><given-names>S</given-names> </name><name name-style="western"><surname>Couture</surname><given-names>B</given-names> </name><name name-style="western"><surname>Kang</surname><given-names>MJ</given-names> </name><etal/></person-group><article-title>Quantifying and visualizing nursing flowsheet documentation burden in acute and critical care</article-title><source>AMIA Annu Symp Proc</source><year>2018</year><volume>2018</volume><fpage>348</fpage><lpage>357</lpage><pub-id pub-id-type="medline">30815074</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sutton</surname><given-names>DE</given-names> </name><name name-style="western"><surname>Fogel</surname><given-names>JR</given-names> </name><name name-style="western"><surname>Giard</surname><given-names>AS</given-names> </name><name name-style="western"><surname>Gulker</surname><given-names>LA</given-names> </name><name name-style="western"><surname>Ivory</surname><given-names>CH</given-names> </name><name name-style="western"><surname>Rosa</surname><given-names>AM</given-names> </name></person-group><article-title>Defining an essential clinical dataset for admission patient history to reduce nursing documentation burden</article-title><source>Appl Clin Inform</source><year>2020</year><month>05</month><volume>11</volume><issue>3</issue><fpage>464</fpage><lpage>473</lpage><pub-id pub-id-type="doi">10.1055/s-0040-1713634</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Madandola</surname><given-names>OO</given-names> </name><name name-style="western"><surname>Bjarnadottir</surname><given-names>RI</given-names> </name><name name-style="western"><surname>Yao</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>The relationship between electronic health records user interface features and data quality of patient clinical information: an integrative review</article-title><source>J Am Med Inform Assoc</source><year>2023</year><month>12</month><day>22</day><volume>31</volume><issue>1</issue><fpage>240</fpage><lpage>255</lpage><pub-id pub-id-type="doi">10.1093/jamia/ocad188</pub-id><pub-id pub-id-type="medline">37740937</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Diaz-Garelli</surname><given-names>F</given-names> </name><name name-style="western"><surname>Strowd</surname><given-names>R</given-names> </name><name name-style="western"><surname>Lawson</surname><given-names>VL</given-names> </name><etal/></person-group><article-title>Workflow differences affect data accuracy in oncologic EHRs: a first step toward detangling the diagnosis data babel</article-title><source>JCO Clin Cancer Inform</source><year>2020</year><month>06</month><volume>4</volume><fpage>529</fpage><lpage>538</lpage><pub-id pub-id-type="doi">10.1200/CCI.19.00114</pub-id><pub-id pub-id-type="medline">32543899</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Johnson</surname><given-names>CM</given-names> </name><name name-style="western"><surname>Johnston</surname><given-names>D</given-names> </name><name name-style="western"><surname>Crowley</surname><given-names>PK</given-names> </name><etal/></person-group><article-title>EHR usability toolkit: a background report on usability and electronic health records</article-title><source>Agency for Healthcare Research and Quality, US Department of Health and Human Services</source><year>2011</year><access-date>2024-08-25</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://digital.ahrq.gov/sites/default/files/docs/citation/EHR_Usability_Toolkit_Background_Report.pdf">https://digital.ahrq.gov/sites/default/files/docs/citation/EHR_Usability_Toolkit_Background_Report.pdf</ext-link></comment></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Staggers</surname><given-names>N</given-names> </name><name name-style="western"><surname>Elias</surname><given-names>BL</given-names> </name><name name-style="western"><surname>Makar</surname><given-names>E</given-names> </name><name name-style="western"><surname>Alexander</surname><given-names>GL</given-names> </name></person-group><article-title>The imperative of solving nurses&#x2019; usability problems with health information technology</article-title><source>J Nurs Adm</source><year>2018</year><month>04</month><volume>48</volume><issue>4</issue><fpage>191</fpage><lpage>196</lpage><pub-id pub-id-type="doi">10.1097/NNA.0000000000000598</pub-id><pub-id pub-id-type="medline">29570144</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Panetta</surname><given-names>K</given-names> </name></person-group><article-title>A data and analytics leader&#x2019;s guide to data literacy</article-title><source>Gartner</source><year>2021</year><access-date>2024-01-13</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.gartner.com/smarterwithgartner/a-data-and-analytics-leaders-guide-to-data-literacy#:~:text=What%20is%20data%20literacy%3F,case%2C%20application%20and%20resulting%20value">https://www.gartner.com/smarterwithgartner/a-data-and-analytics-leaders-guide-to-data-literacy#:~:text=What%20is%20data%20literacy%3F,case%2C%20application%20and%20resulting%20value</ext-link></comment></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Stobierski</surname><given-names>T</given-names> </name></person-group><article-title>Data literacy: an introduction for business</article-title><source>Harvard Business School Online</source><year>2021</year><access-date>2024-01-13</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://online.hbs.edu/blog/post/data-literacy">https://online.hbs.edu/blog/post/data-literacy</ext-link></comment></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Ray</surname><given-names>B</given-names> </name></person-group><article-title>Why marketers need a chief data ethics officer now</article-title><source>AiThority</source><year>2023</year><access-date>2024-08-27</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://aithority.com/technology/why-marketers-need-a-chief-data-ethics-officer-now/">https://aithority.com/technology/why-marketers-need-a-chief-data-ethics-officer-now/</ext-link></comment></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Faraj</surname><given-names>S</given-names> </name><name name-style="western"><surname>Pachidi</surname><given-names>S</given-names> </name><name name-style="western"><surname>Sayegh</surname><given-names>K</given-names> </name></person-group><article-title>Working and organizing in the age of the learning algorithm</article-title><source>Inf Org</source><year>2018</year><month>03</month><volume>28</volume><issue>1</issue><fpage>62</fpage><lpage>70</lpage><pub-id pub-id-type="doi">10.1016/j.infoandorg.2018.02.005</pub-id></nlm-citation></ref></ref-list></back></article>