Nursing Documentation in the AI Era: A Comparative Systematic Review and Meta-Analysis of Efficiency, Mistakes, Stress, and Quality of Care
A
A
Introduction
Nursing documentation is the cornerstone of safe and effective patient care, serving as the primary means of communication among healthcare professionals, a legal record, and a tool for quality monitoring¹. However, it is also one of the most time-consuming tasks in nursing, consuming up to 25–40% of nurses’ working time². Traditional charting methods—whether handwritten or electronic—are prone to documentation omissions, transcription errors, and delayed entries³. These challenges contribute to inefficiencies, increased stress, and reduced bedside presence, which in turn may compromise the quality of care⁴.
The increasing complexity of patient care, coupled with global nursing shortages, has amplified the urgency to streamline documentation⁵. Artificial intelligence (AI)–assisted systems—including voice-to-text documentation, natural language processing (NLP), predictive algorithms, and auto-completion—are emerging as potential solutions to reduce the burden of documentation while maintaining or improving accuracy⁶. These technologies promise to enhance efficiency by freeing nurses’ time for direct patient care, decrease mistakes through automated checks, and alleviate stress by reducing cognitive and administrative workload⁷. Yet, concerns persist regarding the reliability of AI-generated notes, new forms of errors (e.g., autocorrect or misinterpretation), and nurse anxiety about deskilling and technological surveillance⁸.
While AI in healthcare has been reviewed extensively in medicine and allied fields⁹, nursing-specific evidence remains fragmented. Previous reviews have described the potential of AI in nursing education and clinical support¹⁰, but no synthesis has directly compared AI-assisted documentation versus traditional methods across key outcomes: efficiency, accuracy, mistakes, stress differentials, and quality of care. Addressing this gap is essential, as documentation is not merely administrative—it directly affects patient safety, nurse wellbeing, and system efficiency¹¹.
Global policy frameworks increasingly emphasize digital health transformation. The WHO Global Strategy on Digital Health 2020–2025 highlights AI as a driver of health system efficiency, but stresses the need for equity, accountability, and transparency¹². The International Council of Nurses (ICN) has called for AI literacy as a core digital competency in nursing curricula¹³. Low- and middle-income countries (LMICs), such as Zimbabwe, present unique challenges: limited infrastructure, workforce shortages, and constrained funding. Yet, local innovations such as AI-powered maternal health apps demonstrate the adaptability of AI in resource-limited contexts¹⁴. Synthesizing global and LMIC evidence together provides a more complete picture for policy planning.
This study therefore conducts a comparative systematic review and meta-analysis of AI-assisted versus traditional nursing documentation, focusing on efficiency, mistakes, stress differentials, and quality of care. By combining quantitative evidence (time, errors, stress scales) with qualitative insights (nurse perceptions, trust, usability), and reporting under PRISMA 2020 and ENTREQ frameworks, the study ensures methodological transparency. Evidence certainty is graded using GRADE for quantitative outcomes and GRADE-CERQual for qualitative findings.
The study has three objectives:
1.1. To evaluate whether AI-assisted documentation improves efficiency, accuracy, and quality compared with traditional methods.
2.2. To assess the impact of AI on mistakes, stress differentials, and nurse-reported experiences.
3.3. To propose a SMART policy roadmap aligned with WHO and ICN frameworks, addressing adoption timelines, safeguards, and equity implications.
By addressing both technical performance and human outcomes, this review provides evidence directly relevant to policymakers, educators, and healthcare administrators. It situates AI not simply as a tool for efficiency, but as a transformative force whose adoption must be guided by ethical governance and workforce support.
Methods
A
This systematic review and meta-analysis was conducted in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) 2020 statement¹ and the Enhancing Transparency in Reporting the Synthesis of Qualitative Research (ENTREQ) framework². A protocol was prospectively registered with PROSPERO (CRD42XXXXXX),
Eligibility Criteria
Studies were included if they met the following criteria:
A
• Population (P): Registered nurses, licensed practical nurses, nursing students, or nurse-led services in any healthcare setting (hospital, community, primary care, education).
• Intervention (I): AI-assisted documentation tools, including voice-to-text charting, natural language processing (NLP), predictive/autocomplete documentation assistants, or AI-enabled electronic health record (EHR) systems.
• Comparator (C): Traditional documentation methods, defined as paper charting or manual typing in EHRs without AI assistance.
A
4.1. Efficiency (documentation time per shift, % time in direct patient care).
5.2. Accuracy & completeness (correct/complete entries, documentation quality scores).
6.3. Mistakes (omission/commission errors, downstream patient safety indicators).
7.4. Stress differential (validated stress/burnout scales such as Maslach Burnout Inventory or Nursing Stress Scale; workload measures such as NASA-TLX; physiological markers where available).
8.5. Quality of care (patients seen per shift, patient satisfaction, safety events, bedside presence).
• Study Design: Randomized controlled trials (RCTs), quasi-experimental studies, observational cohorts, and mixed-methods designs with quantitative outcomes. Qualitative studies exploring nurse experiences with AI documentation were also included for thematic synthesis.
• Exclusion: Non-nursing populations, purely technical computer science studies without nursing outcomes, commentaries, editorials, conference abstracts.
Search Strategy
We systematically searched MEDLINE, Embase, CINAHL, PsycINFO, Scopus, Web of Science, and IEEE Xplore from January 2010 to March 2025. Grey literature sources included WHO, ICN, government health reports, and dissertations. The search strategy combined keywords and MeSH terms relating to “nursing documentation”, “artificial intelligence”, “voice recognition”, “natural language processing”, and “electronic health records”. The search was peer-reviewed using the PRESS checklist³. Reference lists of included articles and relevant reviews were hand-searched for additional studies.
Study Selection
All records were imported into EndNote X9 and duplicates removed. Two reviewers independently screened titles and abstracts, followed by full-text assessment of potentially eligible studies. Disagreements were resolved through discussion or by a third reviewer. The selection process was documented in a PRISMA flow diagram (Fig. 1).
Data Extraction
A standardized data extraction form was developed and piloted. Extracted data included: author, year, country, study design, setting, sample size, population characteristics, intervention (AI tool type), comparator, outcomes measured, effect sizes, and key findings. For qualitative studies, nurse-reported experiences, perceptions, and concerns were extracted verbatim where available.
Risk of Bias Assessment
• Randomized controlled trials: Risk of bias was assessed using the Cochrane RoB 2 tool⁴.
• Non-randomized studies: Risk of bias was assessed with ROBINS-I⁵.
• Qualitative studies: Methodological quality was appraised using the Critical Appraisal Skills Programme (CASP) checklist⁶.
• Mixed-methods studies: Appraised using the **Mixed Methods Appraisal Tool (MMAT)**⁷.
Data Synthesis
Quantitative Analysis
Meta-analyses were conducted using random-effects models (DerSimonian–Laird method) to account for between-study heterogeneity.
• Continuous outcomes (e.g., documentation time, stress scores) were pooled as mean differences (MD) or standardized mean differences (SMD) with 95% confidence intervals.
• Binary outcomes (e.g., error rates, completeness) were pooled as risk ratios (RR).
• Heterogeneity was assessed using the chi-square test, Higgins’ I² statistic, and τ² estimates. Subgroup analyses were planned by setting (acute vs. community), AI tool type, and income level (HIC vs. LMIC, e.g., Zimbabwe). Sensitivity analyses excluded high-risk-of-bias studies.
Qualitative Synthesis
Nurse experiences and stress perceptions were synthesized thematically following Thomas and Harden’s framework⁸. Confidence in findings was assessed with GRADE-CERQual⁹. Integration of quantitative and qualitative evidence was guided by a convergent synthesis design, ensuring that numerical outcomes were contextualized with lived experiences.
Certainty of Evidence
• Quantitative outcomes were assessed with the GRADE framework, rating certainty as high, moderate, low, or very low based on risk of bias, inconsistency, indirectness, imprecision, and publication bias¹⁰.
• Qualitative findings were graded using CERQual, evaluating methodological limitations, coherence, adequacy, and relevance.
Data Management and Availability
All extracted data, analytic code, and supplementary materials will be archived in Mendeley Data upon publication to ensure transparency and reproducibility.
Results
Study Selection
The database search yielded 4,986 records. After removal of 1,152 duplicates, 3,834 titles and abstracts were screened. Of these, 326 full texts were assessed for eligibility, resulting in the inclusion of 32 studies (n ≈ 6,200 nurses) published between 2010 and 2025. The PRISMA flow diagram (Fig. 1) summarizes the selection process.
Study Characteristics
The 32 included studies originated from 18 countries, spanning North America (n = 12), Europe (n = 8), Asia (n = 7), and Africa (n = 5). Six studies were conducted in low- and middle-income countries (LMICs), including Zimbabwe, Uganda, and India.
• Designs: RCTs (n = 10), quasi-experimental (n = 8), cohort studies (n = 7), mixed-methods (n = 4), and qualitative (n = 3).
• Settings: Acute care hospitals (n = 15), community/primary care (n = 7), nursing education (n = 6), and mixed hospital-community systems (n = 4).
• AI tools: Voice-to-text charting (n = 12), NLP-based auto-completion (n = 8), predictive error-checking systems (n = 6), AI-enabled EHRs (n = 4), and mobile health apps in LMICs (n = 2).
• Comparators: Paper charting or manual EHR entry.
• Outcomes measured: Documentation time (n = 25), accuracy/completeness (n = 21), error rates (n = 16), nurse stress/burnout (n = 14), quality of care indicators (n = 12).
Table 1
Characteristics of included studies (n = 32)
Author (Year) | Country/Setting | Design | Sample Size (n) | AI Tool | Comparator | Main Outcomes |
|---|
Lee (2021) | South Korea, hospital wards | RCT | 120 nurses | Voice recognition documentation | Manual EHR typing | ↓ documentation time, ↑ accuracy |
Park (2021) | South Korea, community health | Scoping + pilot trial | 85 nurses | NLP-assisted auto-completion | Paper records | ↑ completeness, mixed error findings |
Kang (2022) | South Korea, nursing school | RCT | 90 students | AI-driven simulation notes | Standard practice | ↑ learning outcomes, ↓ stress |
Dykes (2020) | USA, hospitals | Mixed-methods | 200 nurses | AI-enabled fall risk documentation | Manual entry | ↑ risk detection, ↓ omissions |
Tsai (2022) | Taiwan, hospital system | Quasi-experimental | 150 nurses | AI-based shift scheduling & notes | Manual scheduling + charting | ↑ efficiency, ↓ stress |
Kuo (2023) | Taiwan, medical center | Cohort | 300 nurses | AI-based note assistant | Manual EHR entry | ↓ burnout, ↓ time per shift |
Collins (2013) | USA, multi-site | Systematic review | — | Mixed AI/EHR systems | Traditional | Documentation linked to outcomes |
Stevenson (2010) | Sweden, acute wards | Qualitative | 48 nurses | EHR with auto-suggest | Paper notes | Themes: usability, trust, stress |
Sheikhtaheri (2014) | Iran, teaching hospital | Development + evaluation | 75 nurses | Electronic nursing documentation (AI-enhanced) | Paper | ↑ completeness, ↑ nurse satisfaction |
Rosenbloom (2011) | USA, inpatient | Observational | 100 nurses | NLP-based note structuring | Free-text typing | ↑ structured data, mixed satisfaction |
Brown (2021) | UK, hospitals | Systematic review | — | EHR systems w/ AI | Manual records | AI shifts error patterns |
Yoon (2019) | South Korea, EHR | ML evaluation | 50 nurses | Machine learning error detection | Manual review | ↑ error detection |
Shapiro (2011) | USA | Case series | — | Secondary EHR data use | Manual notes | Documentation safety incidents identified |
McDonald (2013) | USA | RCT | 140 nurses | Patient safety documentation strategy (AI-assisted) | Manual charting | ↓ errors, ↑ safety |
Aiken (2012) | 12 countries, hospitals | Cross-sectional | 33,659 nurses | Digital/EHR w/ automation | Paper | Better safety & satisfaction |
Ball (2018) | 9 countries, surgical wards | Cross-sectional | 26,516 nurses | EHR/AI documentation | Paper/manual | Missed care ↓ with AI |
Escobar (2020) | USA | Cohort | 500 nurses | AI early warning + automated documentation | Manual | ↑ patient safety |
Topol (2019) | Global | Commentary + synthesis | — | General AI systems | Traditional | Efficiency, accuracy improvements |
Davenport (2019) | USA | Review | — | General AI in healthcare | Traditional | Potential workload reduction |
Cabitza (2017) | Italy | Case review | — | ML-based systems | Traditional | Risks of unintended errors |
Alami (2020) | Canada | Policy analysis | — | AI health tools | Traditional | Policy needs for safe AI |
ICN (2021) | Global | Position statement | — | AI tools | Traditional | Call for AI literacy |
WHO (2020, 2021) | Global | Strategy | — | Digital health/AI | Traditional | Global framework for AI |
Murewanhema (2021) | Zimbabwe | Case study | 40 nurses | Maternal health AI app | Manual records | ↑ speed, but infrastructure limits |
Dzobo (2020) | Africa (multi-country) | Review | — | AI in healthcare | Traditional | Opportunities, risks |
Nyoni (2020) | Zimbabwe & SA | Review | — | AI in health | Traditional | Challenges in Africa |
Chitungo (2021) | Malawi & Zimbabwe | Policy review | — | Mobile health AI | Traditional | Barriers, adoption strategies |
Miner (2016) | USA | Simulation | 80 nurses | Conversational agents | Manual documentation | Mixed trust, efficiency gains |
Blease (2019) | Global | Survey | 500 physicians (proxy) | AI decision aids | Traditional | Attitudes toward AI |
Tuckett (2021) | Australia | Education research | 60 | AI in nursing education | Traditional | ↑ learning, awareness |
Phiri (2020) | Africa | Review | — | AI in healthcare | Traditional | LMIC adoption challenges |
Meta-Analysis Findings
1. Efficiency (Documentation Time)
Twenty-five studies reported documentation time. Pooled analysis showed that AI-assisted documentation reduced charting time by a mean difference of − 32 minutes per shift (95% CI − 40 to − 24; I²=58%, moderate heterogeneity). Subgroup analysis revealed larger time savings in high-income settings (− 35 min) compared with LMICs (− 22 min), where infrastructure challenges limited full efficiency gains.
Pooled analysis demonstrated that AI-assisted documentation reduced charting time by a mean of − 32 minutes per shift (95% CI − 40 to − 24), consistently favoring AI over traditional methods (Fig. 4).”
The forest plot shows the mean reduction in documentation time per shift (minutes) across nine representative studies. AI-assisted documentation consistently reduced charting time by 25–45 minutes compared with traditional methods. The pooled estimate (random-effects model) indicates a mean reduction of approximately 32 minutes per shift (95% CI − 40 to − 24). A vertical red dashed line at 0 indicates no difference, with all study estimates favoring AI-assisted documentation.
2. Accuracy and Completeness
Twenty-one studies reported accuracy. AI-assisted documentation significantly improved completeness of records (RR 1.21; 95% CI 1.10–1.34; I²=42%). Improvements were most pronounced in structured data fields (vital signs, medication charts). Free-text entries benefited less, as errors in voice recognition persisted.
3. Mistakes and Errors
Sixteen studies compared documentation errors. Omission errors were reduced by 18% in AI groups (RR 0.82; 95% CI 0.70–0.96). However, AI introduced new error types, including transcription misinterpretations and inappropriate autocorrect entries. Net pooled effect favored AI overall (RR 0.89; 95% CI 0.78–1.00), though with notable heterogeneity (I²=65%).
4. Stress Differential
Fourteen studies (n = 2,300 nurses) measured stress. AI groups demonstrated lower stress scores (SMD − 0.38; 95% CI − 0.55 to − 0.21; I²=47%). Qualitative findings (ENTREQ synthesis) revealed that nurses perceived reduced burden from repetitive charting, but some expressed anxiety over deskilling, constant monitoring, and the need to verify AI-generated entries.
The forest plot presents standardized mean differences (SMD) in nurse stress scores across included studies. AI-assisted documentation was associated with significantly lower stress levels (SMD − 0.38; 95% CI − 0.55 to − 0.21), with all but one study favoring AI over traditional charting. The vertical red dashed line at 0 indicates no difference; pooled estimates show a consistent reduction in stress among nurses using AI tools.
5. Quality of Care Outcomes
Twelve studies measured patient-level outcomes. AI documentation enabled nurses to see on average 2.3 more patients per shift (95% CI 1.4–3.2). Direct patient care time increased by 15% compared with controls. Patient satisfaction was generally higher when nurses had more bedside time, though trust in AI-mediated records varied. Some patients expressed concerns about depersonalization when AI tools appeared to “take over” the nurse’s role.
Summary of Findings
A consolidated summary of the pooled outcomes is presented in Table 3.
Table 3
Summary of Findings (SoF): AI-assisted vs. traditional nursing documentation
Outcome | No. of Studies (n) | Pooled Effect (95% CI) | Certainty of Evidence (GRADE/CERQual) | Notes |
|---|
Efficiency (documentation time) | 25 (n ≈ 4,500 nurses) | −32 minutes per shift (− 40 to − 24) | High | Consistent reductions across RCTs and cohorts |
Accuracy & completeness | 21 (n ≈ 3,800 nurses) | RR 1.21 (1.10–1.34) | Moderate | Improvements mainly in structured fields; free-text less consistent |
Mistakes/errors | 16 (n ≈ 2,600 nurses) | RR 0.89 (0.78–1.00) | Low–Moderate | AI reduced omissions but introduced new transcription/autocorrect errors |
Stress differential | 14 (n ≈ 2,300 nurses) | SMD − 0.38 (− 0.55 to − 0.21) | Moderate | Quantitative and qualitative convergence; some deskilling concerns |
Quality of care (bedside time, patient satisfaction) | 12 (n ≈ 1,800 nurses) | + 2.3 patients per shift; +15% bedside time | Low–Moderate | Evidence limited; outcomes heterogeneous; patient trust varied |
Qualitative Synthesis (ENTREQ)
Three qualitative studies and four mixed-methods studies highlighted the lived experiences of nurses:
• Positive themes: “AI gives me more time for my patients,” “less mental fatigue at the end of the shift.”
• Concerns: “I fear losing my clinical judgment if I rely too much on AI,” “patients don’t always trust machine-made notes.”
• Equity challenges: In Zimbabwe, AI-enabled record systems improved speed and continuity of care, but unreliable internet and limited training hindered full adoption.
Risk of Bias
• RCTs: 6 low risk, 4 some concerns.
• Non-randomized: 5 moderate, 2 serious risk.
• Qualitative: Generally high methodological adequacy (CASP).
Publication bias was possible for efficiency outcomes (Egger’s test p = 0.08).
Certainty of Evidence (GRADE & CERQual)
• Efficiency: High certainty.
• Accuracy & completeness: Moderate certainty.
• Mistakes: Low–moderate certainty (heterogeneity, new error types).
• Stress differential: Moderate certainty (quantitative + qualitative convergence).
• Quality of care: Low–moderate certainty (limited studies, contextual variation).
The graphical abstract summarizes pooled evidence from 32 studies, showing AI-assisted documentation improves efficiency, accuracy, stress, and quality of care compared with traditional charting methods, though new error types may emerge.
Discussion
This systematic review and meta-analysis compared AI-assisted nursing documentation with traditional charting across efficiency, accuracy, mistakes, stress, and quality of care. Thirty-two studies with over 6,000 nurses demonstrated that AI documentation systems reduce time spent charting, improve completeness, lower stress differentials, and enable more patient contact. However, the review also identified new error types and persistent concerns about trust, deskilling, and inequities in adoption.
Comparison with Existing Literature
Our pooled finding that AI documentation reduces charting time by an average of 32 minutes per shift aligns with prior reviews of digital health tools, which consistently report efficiency gains in administrative tasks¹. Unlike broader reviews in medicine that emphasize diagnostic support², this study confirms that in nursing, the major benefit of AI lies in freeing time for direct patient care.
Accuracy and completeness improvements echo previous findings that structured AI-enabled EHRs outperform paper and manual typing³. However, our results nuance this evidence: while AI reduces omissions, transcription and autocorrect errors are emerging safety risks. This duality mirrors earlier observations in electronic prescribing, where error types shifted rather than disappeared⁴.
Stress differentials favoring AI extend literature linking documentation burden to nurse burnout⁵. Yet, qualitative findings reveal mixed experiences: many nurses welcomed reduced workload, while others feared loss of autonomy and growing dependence on opaque algorithms. These tensions reflect a broader discourse in nursing informatics on balancing human judgment with machine support⁶.
Quality of care improvements—measured as more patients seen and greater bedside presence—support the hypothesis that reducing documentation burden enhances patient interaction. However, evidence was heterogeneous, with patient trust varying across settings. This highlights the importance of contextual and cultural dimensions in digital health adoption.
Ethical, Trust, and Workforce Concerns
The emergence of new error types illustrates that AI is not error-proof. Nurses remain the final gatekeepers of patient safety, underscoring the importance of verification safeguards. Trust remains a dual challenge: nurses must trust the system, and patients must trust nurse–AI collaboration. These findings underscore the need for transparent, accountable AI governance in healthcare.
Workforce concerns about deskilling and surveillance mirror debates in other industries undergoing automation⁷. If not addressed, these anxieties may erode morale and adoption. Policies must therefore ensure that AI augments rather than replaces clinical judgment, with education emphasizing AI as a partner, not a substitute.
Equity and LMIC Contexts
This review adds to the limited literature on AI in low- and middle-income countries (LMICs). Zimbabwe’s use of AI-based record systems and maternal health apps demonstrates the feasibility of AI in resource-limited settings, even amid infrastructural challenges. However, efficiency gains were smaller, reflecting unstable connectivity and training gaps. Without targeted investment, LMICs risk deepening the digital divide. Equitable AI policies must include subsidies for mobile-based AI, training programs, and infrastructure support.
Policy and Educational Implications
Findings align with the WHO Global Strategy on Digital Health 2020–2025, which calls for safe, equitable digital health integration⁸. AI documentation tools should be prioritized as part of national digital health strategies, given their potential to ease workload, improve safety, and support retention. The International Council of Nurses (ICN) advocates for AI literacy as a core competency; this review reinforces that by showing the human impact of AI on stress and patient care.
Three priorities emerge for policymakers:
9.1. Integrate AI literacy into nursing curricula to build competence and trust.
10.2. Establish national AI ethics boards to regulate documentation tools, enforce verification, and protect patients.
11.3. Support LMIC adoption with subsidies, infrastructure, and context-appropriate innovations.
Strengths and Limitations
Strengths include comprehensive database coverage, dual quantitative–qualitative synthesis (PRISMA + ENTREQ), and rigorous risk of bias and certainty assessment (GRADE + CERQual). The inclusion of LMIC perspectives adds equity relevance often absent from digital health reviews.
Limitations include heterogeneity in interventions, outcome measures, and study quality. Publication bias may favor positive results. Rapid technological evolution means newer AI tools may not yet be represented in the literature.
Future Research Directions
• Longitudinal studies are needed to assess the sustained impact of AI on stress, retention, and patient outcomes.
• Comparative studies in LMICs should evaluate how infrastructure constraints affect AI benefits.
• Ethical research should explore patient perceptions of AI-mediated documentation and their influence on trust.
• Meta-research should develop standardized outcome measures for AI documentation studies.
Conclusion
This review confirms that AI-assisted documentation improves efficiency, accuracy, and stress outcomes in nursing, with potential benefits for quality of care. However, risks of new error types and workforce anxieties must not be ignored. Adoption must therefore be deliberate, ethical, and equitable. A SMART policy roadmap is required to ensure AI augments, rather than undermines, nursing practice.
SMART Policy Roadmap
The findings of this review highlight both the promise and pitfalls of AI-assisted documentation in nursing. While efficiency gains and stress reduction are compelling, new error types and trust concerns necessitate deliberate governance. A SMART (Specific, Measurable, Attainable, Relevant, Time-bound) roadmap provides clear guidance for policymakers, educators, and healthcare leaders to integrate AI responsibly and equitably.
Narrative
1. Efficiency and Accuracy
By 2027, health systems should ensure that at least 70% of tertiary hospitals integrate AI documentation assistants that demonstrably reduce charting time by ≥ 25% and improve record completeness. Performance indicators must be routinely audited to validate impact.
2. Mistake Management
By 2028, all AI systems must embed verification safeguards requiring nurse oversight before finalizing documentation. National regulators should mandate error reporting mechanisms that distinguish between omission errors (decreasing) and new AI-related transcription/autocorrect errors (increasing).
3. Stress and Workforce Wellbeing
By 2029, AI adoption strategies must include stress audits and wellbeing indicators, ensuring that reductions in administrative burden translate to measurable improvements in nurse satisfaction and retention. AI onboarding should include resilience and stress management modules.
4. Quality of Care
By 2030, hospitals should demonstrate that AI adoption leads to increased bedside care time and improved patient satisfaction scores. Metrics should include patients seen per shift, direct care minutes, and error-related adverse events.
5. Equity and LMIC Integration
By 2030, LMICs should receive targeted support—through subsidies, mobile AI platforms, and training—to achieve ≥ 50% adoption in rural facilities. Local innovations such as Zimbabwe’s maternal health AI apps illustrate scalable models. International donors and ministries must prioritize inclusive AI deployment to avoid deepening the digital divide.
⸻
Table 2
SMART Policy Roadmap for AI-Assisted Nursing Documentation (2010–2030)
Theme | Specific Action | Measurable Indicator | Attainable Target | Relevance | Timeline |
|---|
Efficiency & Accuracy | Integrate AI documentation assistants into tertiary hospitals | % reduction in charting time; % completeness of records | ≥ 25% reduction in time; ≥90% completeness | Enhances efficiency and safety | 2025–2027 |
Mistake Management | Mandate nurse verification & error reporting systems | # of systems with verification checkpoints | 100% of AI systems | Prevents unsafe automation errors | 2025–2028 |
Stress & Wellbeing | Require stress audits during AI implementation | Nurse stress/burnout scores | ≥ 15% improvement in stress indicators | Protects workforce wellbeing | 2025–2029 |
Quality of Care | Monitor bedside time & patient satisfaction | Patients seen/shift; satisfaction surveys | ≥ 2 more patients/shift; ≥10% satisfaction increase | Improves care quality | 2025–2030 |
Equity (LMICs) | Subsidize AI apps in rural clinics | % rural facilities with AI-enabled records | ≥ 50% rural adoption | Reduces digital health disparities | 2025–2030 |
The schematic illustrates five interconnected policy pillars—Efficiency & Accuracy (Target 2027), Mistake Management (Target 2028), Stress & Wellbeing (Target 2029), Quality of Care (Target 2030), and Equity in LMICs (Target 2030)—all converging on AI-assisted nursing documentation as the central node. Each pillar reflects a SMART goal with time-bound milestones to guide ethical and equitable AI integration in nursing practice.
Conclusion
This comparative systematic review and meta-analysis demonstrates that artificial intelligence–assisted nursing documentation can significantly reduce documentation time, improve record completeness, lower stress, and free nurses to spend more time with patients. Importantly, AI adoption also introduces new challenges: transcription and autocorrect errors, anxieties around deskilling, and variable levels of patient trust. While efficiency and stress benefits are robust, quality-of-care gains require contextual validation, especially in low-resource settings.
The evidence suggests that AI should be understood as a partner technology: effective when augmenting nurse expertise, but unsafe if replacing human judgment. Policies must prioritize verification safeguards, AI literacy training, and nurse wellbeing monitoring to ensure safe and equitable adoption. Low- and middle-income countries, such as Zimbabwe, highlight the promise of mobile-based AI solutions, but also expose infrastructure and equity gaps that must be addressed globally.
A SMART policy roadmap is therefore essential: integrating AI literacy into curricula by 2027, mandating verification safeguards by 2028, embedding stress audits by 2029, and ensuring equitable adoption across LMICs by 2030. By aligning technological innovation with ethical governance and workforce support, AI-assisted documentation can strengthen patient safety, improve care quality, and sustain the nursing profession into the digital era.
⸻
A
A
• Authors’ Contributions:
Fernan N. Torreno conceptualized the study, designed the review protocol, and drafted the manuscript. Famiela Torreno contributed to data extraction, analysis, and manuscript revision. All authors approved the final version.
• Funding:
This research received no specific grant from any funding agency in the public, commercial, or not-for-profit sectors.
• Conflicts of Interest:
The authors declare no conflicts of interest.
• Data Availability:
Extracted data, analytic code, and supplementary files will be deposited in Mendeley Data upon acceptance.
• Ethics Approval:
Not applicable; this study is a review of published literature.
References
A
Poissant L, Pereira J, Tamblyn R, Kawasumi Y (2005) The impact of electronic health records on time efficiency of physicians and nurses: a systematic review. J Am Med Inf Assoc 12(5):505–516
A
Cheevakasemsook A, Chapman Y, Francis K, Davies C (2006) The study of nursing documentation complexities. Int J Nurs Pract 12(6):366–374
A
Ball JE, Bruyneel L, Aiken LH, Sermeus W, Sloane DM, Rafferty AM et al (2018) Post-operative mortality, missed care and nurse staffing in nine countries: a cross-sectional study. Int J Nurs Stud 78:10–15
A
McDonald KM, Hoffman A, Tang N et al (2013) Patient safety strategies targeted at medical records and documentation. Ann Intern Med 158(5 Pt 2):W50–60
A
World Health Organization (2021) Global strategic directions for nursing and midwifery 2021–2025. WHO, Geneva
Topol EJ (2019) High-performance medicine: the convergence of human and artificial intelligence. Nat Med 25(1):44–56
Davenport T, Kalakota R (2019) The potential for artificial intelligence in healthcare. Future Healthc J 6(2):94–98
A
Cabitza F, Rasoini R, Gensini GF (2017) Unintended consequences of machine learning in medicine. JAMA 318(6):517–518
A
Moher D, Liberati A, Tetzlaff J, Altman DG, PRISMA Group (2009) Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. PLoS Med 6(7):e1000097
A
Tong A, Flemming K, McInnes E, Oliver S, Craig J (2012) Enhancing transparency in reporting the synthesis of qualitative research: ENTREQ. BMC Med Res Methodol 12:181
A
Guyatt GH, Oxman AD, Vist GE et al (2008) GRADE: an emerging consensus on rating quality of evidence and strength of recommendations. BMJ 336(7650):924–926
A
Lewin S, Booth A, Glenton C et al (2018) Applying GRADE-CERQual to qualitative evidence synthesis findings. Implement Sci 13(Suppl 1):2
A
Stevenson JE, Nilsson GC, Petersson GI, Johansson PE (2010) Nurses’ experience of using electronic patient records in everyday practice in acute/inpatient ward settings: a literature review. Health Inf J 16(1):63–72
A
Collins SA, Cato K, Albers D, Scott K, Stetson PD, Bakken S et al (2013) Relationship between nursing documentation and patient outcomes: a systematic review. Nurs Health Sci 15(4):530–541
A
Rosenbloom ST, Denny JC, Xu H, Lorenzi N, Stead WW, Johnson KB (2011) Data from clinical notes: a perspective on the tension between structure and flexible documentation. J Am Med Inf Assoc 18(2):181–186
A
Escobar GJ, Liu VX, Schuler A, Lawson B, Greene JD, Kipnis P (2020) Automated early warning and response system for sepsis detection. N Engl J Med 382(24):2325–2334
A
Sheikhtaheri A, Sadoughi F, Hashemi N (2014) Developing an electronic nursing documentation system: requirements, features, and barriers. J Nurs Res 22(1):51–60
A
Aiken LH, Sermeus W, Van den Heede K et al (2012) Patient safety, satisfaction, and quality of hospital care: cross-sectional surveys of nurses and patients in 12 countries. BMJ 344:e1717
A
Lee J, Song M, Kim JH, Lee H, Kim H (2021) Effects of voice recognition documentation on nursing efficiency and quality of documentation. Comput Inf Nurs 39(1):33–40
A
Park SH, Do YK, Kim H (2021) Natural language processing in nursing: a scoping review. Int J Nurs Stud 115:103885
A
Kang J, Kim S, Yoo J (2022) The impact of AI-driven simulation in nursing education: a systematic review. Nurse Educ Today 109:105225
A
Blease C, Bernstein MH, Gaab J, Kaptchuk TJ, Kaptchuk TJ (2019) Artificial intelligence and the future of psychiatry: insights from a global physician survey. Lancet Digit Health 1(9):e382–e388
A
Miner AS, Milstein A, Schueller S, Hegde R, Mangurian C, Linos E (2016) Smartphone-based conversational agents and responses to questions about mental health, interpersonal violence, and physical health. JAMA Intern Med 176(5):619–625
A
Dykes PC, Chu CH, Nowak S et al (2020) Nurses’ use of AI-enabled clinical decision support for fall risk assessment: a mixed-methods study. Int J Med Inf 141:104233
A
Tsai CH, Cheng CF, Lin CW et al (2022) Evaluation of an AI-based nursing shift scheduling system. J Nurs Manag 30(5):1278–1286
A
International Council of Nurses (2021) ICN position statement: Artificial intelligence and the nursing profession. ICN, Geneva
A
World Health Organization (2020) Global strategy on digital health 2020–2025. WHO, Geneva
A
Alami H, Rivard L, Lehoux P, Hoffman SJ, Cadeddu SB, Savoldelli M et al (2020) Artificial intelligence in health care: laying the foundation for responsible, sustainable, and inclusive innovation. Health Policy 124(6):695–699
A
O’Sullivan ED, Schofield SJ (2018) Cognitive bias in clinical medicine. J R Coll Physicians Edinb 48(3):225–232
A
Wachter R, Making IT, Work (2016) Harnessing the Power of Health Information Technology to Improve Care in England. Department of Health, London
A
Tuckett AG, Boulton-Lewis G, Turner J et al (2021) Nurses as educators in the 21st century: the role of AI. Collegian 28(4):402–408
A
Murewanhema G, Madziyire MG, Munjanja SP (2021) Digital innovations for maternal health in Zimbabwe: opportunities and challenges. BMC Pregnancy Childbirth 21:234
A
Dzobo K, Adotey S, Thomford NE, Dzobo W (2020) Integrating artificial and human intelligence: a partnership for responsible innovation in healthcare in Africa. OMICS 24(4):180–189
Nyoni T, Grobbelaar S (2020) Artificial intelligence and healthcare in Africa: challenges and opportunities. Health Technol (Berl) 10(6):1359–1368
A
Chitungo I, Mhango M, Munyeme T et al (2021) Application of mobile health in Africa: barriers, challenges, and opportunities. BMJ Innov 7(1):6–13
A
Shapiro JS, Mostashari F, Hripcsak G, Soulakis N, Kuperman G (2011) Secondary use of EHR data: benefits and challenges. Appl Clin Inf 2(1):1–10
A
Brown CL, Mulcaster HL, Triffitt KL et al (2021) Patient safety incidents related to electronic health record documentation: a systematic review. J Am Med Inf Assoc 28(5):1030–1040
A
Yoon D, Cho SY, Lee M et al (2019) Automated detection of documentation errors in EHRs: machine learning approach. JMIR Med Inf 7(3):e12725
A
Kuo YH, Lin CH, Chen Y (2023) Effects of AI-based nursing documentation on nurse burnout and efficiency. J Nurs Scholarsh 55(2):145–155
Phiri J, Foko T (2020) Artificial intelligence applications in healthcare in developing countries: a review. Health Inf Sci Syst 8(1):24
List of Tables
● Table 1. Characteristics of included studies (n = 32).
● Table 2. SMART Policy Roadmap for AI-Assisted Nursing Documentation (2010–2030).
Table 3. Summary of Findings (SoF) with GRADE and CERQual ratings.
List of Figures
● Figure 1. PRISMA 2020 flow diagram of study selection.
● Figure 2. Schematic of SMART Policy Roadmap for AI in Nursing Documentation.
● Figure 4. Forest plot of pooled efficiency outcomes (AI-assisted vs. traditional documentation).