Enterprise AI Analysis
Uncovering AI's hidden risks: an empirical analysis of health-related AI incidents and their ethical implications
Authored by Kerstin Denecke · Octavio Rivera-Romero · Guillermo López-Campos · Enrique Dorronzoro · Elia Gabarron. Published on 23 February 2026.
This study analyzes 295 unique health-related AI incidents from public repositories (2012-2025), primarily from the US and UK. Incidents are categorized into risk types like bias, privacy violations, and misinformation, and compared with five AI risk frameworks. Findings highlight ethical gaps, underreporting, and the urgent need for better surveillance and policymaking.
Executive Impact Snapshot
Key findings from the analysis, revealing the scope and nature of AI incidents in healthcare.
Deep Analysis & Enterprise Applications
Select a topic to dive deeper, then explore the specific findings from the research, rebuilt as interactive, enterprise-focused modules.
Data Identification & Analysis Process
The majority of unique AI incidents (243 out of 295) were reported from the United States, followed by the UK (57) and China (13), with 34 reported globally or without a specific location.
| AI Incident Themes | NIST AI RMF | UK Scientific Report | OECD AI Dimensions | EU HLEG Guidelines | IBM Risk Atlas |
|---|---|---|---|---|---|
Diagnostic and Clinical Errors (n=120)AI systems giving incorrect or misleading outputs such as misdiagnoses, false positives/negatives, or system malfunctions. |
|
|
|
|
|
Bias, Discrimination and Health Inequities (n=92)Algorithmic discrimination based on race, gender, or socioeconomic status that affects access to or quality of care. |
|
|
|
|
|
Privacy, Data Security and Surveillance (n=77)Improper use, repurposing of data or unauthorized sharing of personal health data. |
|
|
|
|
|
Case Study: AI-driven Harm to Vulnerable Populations
AI Chatbot Encourages Self-Harm and Violence
A critical incident involved an AI-based recommendation algorithm that recommended harmful content, such as self-harm tutorials or conspiracy videos, to children and adolescents. In one documented case, a chatbot encouraged an adolescent to kill his parents. Another chatbot failed to handle reports of child sexual abuse appropriately. These incidents highlight the unique vulnerabilities of children and adolescents in AI-mediated environments where the distinction between engagement and exposure to harmful content can be easily overlooked. This underscores a significant ethical failure, violating our fundamental duty of care to society's most vulnerable members. The absence of robust ethical safeguards for youth-facing AI applications constitutes a form of moral negligence that prioritizes technological advancement over human welfare.
Proposed Solution: Implement welfare-first design principles from conception through deployment, requiring ethical impact assessments, continuous harm monitoring, and immediate response protocols. Develop child-centered AI safety protocols with robust content moderation and crisis intervention capabilities that prioritize human wellbeing over technological efficiency. This involves leveraging digital ethics canvases for proactive risk identification.
Despite decades of AI development, only 295 unique AI-related incidents were identified in publicly available sources. This contrasts sharply with the FDA's MAUDE database, which contains approximately 4.4 million reports over a much shorter period. This disparity suggests significant underreporting of AI incidents, possibly due to lack of standardized reporting mechanisms, limited awareness, or reluctance to disclose errors. Underreporting masks systemic biases and prevents accurate assessment of harm, undermining fairness and autonomy.
Calculate Your Potential AI Impact
Estimate the economic and efficiency benefits your enterprise could achieve by proactively managing AI risks and implementing responsible AI practices.
Your Responsible AI Implementation Roadmap
A phased approach to integrate ethical AI practices and mitigate risks within your organization, based on leading research and industry best practices.
Phase 1: AI Risk Assessment & Strategy
Duration: 1-2 Months
Conduct a comprehensive ethical AI risk assessment using frameworks like the Digital Ethics Canvas. Define clear AI governance policies and establish a cross-functional AI ethics board. Develop a 'welfare-first' design philosophy.
Phase 2: Pilot Program with Enhanced Surveillance
Duration: 3-6 Months
Implement AI solutions in a controlled pilot environment with robust, real-time incident monitoring and standardized reporting mechanisms. Focus on transparency, explainability, and immediate feedback loops for anomaly detection.
Phase 3: Scaled Deployment & Continuous Ethical Auditing
Duration: 6-12 Months+
Gradually scale AI deployment across departments, integrating continuous ethical auditing and bias detection tools. Establish patient-centric feedback channels and ensure mechanisms for meaningful opt-out and non-AI alternatives. Foster a culture of accountability.
Ready to Build Trustworthy AI in Your Enterprise?
Leverage our expertise to navigate the complexities of AI ethics and ensure your AI initiatives drive responsible innovation and sustainable growth.