Submitted:
23 February 2026
Posted:
28 February 2026
You are already at the latest version
Abstract
Keywords:
1. Introduction
2. Theoretical Foundations and Governance Evolution
2.1. Defining Artificial Intelligence in Organizational Contexts
2.2. Evolution of AI Governance Discourse
2.3. Risk Management Theoretical Frameworks
2.4. The NIST AI Risk Management Framework
3. Critical Analysis of AI-Related Risks
3.1. Taxonomic Approach to Risk Classification
| Risk Category | Risk Subcategory | Specific Risk Types | Description and Manifestation | Potential Organizational Impact | Likelihood Factors | Detection Difficulty | Illustrative Examples |
|---|---|---|---|---|---|---|---|
| Technical Risks | Algorithmic | Bias and discrimination | Systematic unfairness in AI outputs affecting protected groups | Legal liability; reputational damage; stakeholder harm | Training data imbalances; proxy variables; historical patterns | Medium-High | Credit scoring disparities; hiring algorithm discrimination |
| Accuracy degradation | Declining model performance over time due to data drift or concept drift | Operational failures; poor decisions; safety incidents | Environmental changes; evolving user behavior; data quality issues | Medium | Fraud detection missing new patterns; diagnostic accuracy decline | ||
| Opacity and inexplicability | Inability to understand or explain AI decision-making processes | Regulatory non-compliance; accountability gaps; user distrust | Model complexity; deep learning architectures; proprietary systems | High | Black-box medical diagnoses; unexplainable credit denials | ||
| Robustness failures | System brittleness when encountering novel or adversarial inputs | Unpredictable behavior; exploitation vulnerability; safety risks | Limited training data diversity; insufficient stress testing | High | Autonomous vehicle edge cases; image recognition failures | ||
| Data-Related | Data quality issues | Errors, incompleteness, or inconsistencies in training or operational data | Model unreliability; biased outputs; operational errors | Poor data governance; integration challenges; legacy systems | Medium | Customer segmentation errors; inventory prediction failures | |
| Privacy violations | Unauthorized collection, use, or inference of personal information | Regulatory penalties; reputational harm; individual harm | Insufficient anonymization; data aggregation; inference attacks | Medium-High | Re-identification from anonymized data; behavioral profiling | ||
| Data poisoning | Malicious manipulation of training data to corrupt model behavior | Compromised system integrity; manipulated outputs | Inadequate data provenance; insufficient validation; insider threats | High | Manipulated recommendation systems; corrupted fraud models | ||
| Security | Adversarial attacks | Deliberate inputs designed to deceive or manipulate AI systems | System exploitation; incorrect outputs; safety compromises | Model accessibility; limited adversarial training; known vulnerabilities | High | Image perturbation attacks; voice spoofing | |
| Model extraction | Unauthorized replication of proprietary AI models through query access | Intellectual property theft; competitive disadvantage | Excessive API access; insufficient monitoring | High | Reverse engineering of commercial models | ||
| Infrastructure vulnerabilities | Security weaknesses in AI system deployment and operation | Data breaches; system compromise; operational disruption | Complex technology stacks; rapid deployment; inadequate security testing | Medium | Cloud configuration errors; API vulnerabilities | ||
| Organizational Risks | Operational | Integration failures | Difficulties incorporating AI into existing workflows and systems | Project delays; cost overruns; abandoned initiatives | Legacy system complexity; inadequate change management | Low-Medium | ERP integration challenges; workflow disruption |
| Skill gaps | Insufficient organizational capabilities to develop, deploy, or oversee AI | Implementation failures; vendor dependence; governance gaps | Talent scarcity; inadequate training; rapid technology evolution | Low | Inability to audit models; poor vendor oversight | ||
| Vendor dependence | Over-reliance on external AI providers with limited organizational control | Continuity risks; cost escalation; reduced customization | Outsourcing strategies; proprietary solutions; limited internal capacity | Low | Provider discontinuation; pricing changes; feature limitations | ||
| Strategic | Automation displacement | Workforce disruption from AI-enabled automation of human tasks | Employee relations issues; knowledge loss; transition costs | Aggressive automation targets; inadequate transition planning | Low | Customer service automation; manufacturing robotics | |
| Deskilling | Erosion of human expertise through over-reliance on AI assistance | Capability atrophy; reduced human judgment; succession risks | Extended AI dependence; reduced human practice; knowledge management gaps | Medium | Diagnostic skill erosion in radiology; reduced analytical capabilities | ||
| Competitive disruption | Strategic risks from AI-enabled market changes or competitor advantages | Market share loss; business model obsolescence | Industry dynamics; technology adoption patterns | Medium | Fintech disruption of traditional banking; AI-native competitors | ||
| Governance | Accountability gaps | Unclear responsibility allocation for AI-related decisions and outcomes | Legal exposure; governance failures; stakeholder harm | Distributed development; complex systems; inadequate policies | Medium | Unclear liability for autonomous decisions | |
| Oversight failures | Inadequate monitoring and control mechanisms for AI system behavior | Undetected problems; compliance violations; harm accumulation | Resource constraints; technical complexity; monitoring gaps | Medium-High | Undetected model drift; unmonitored bias emergence | ||
| Societal Risks | Ethical | Autonomy infringement | AI systems that manipulate, deceive, or unduly influence human decision-making | Individual harm; trust erosion; regulatory intervention | Persuasive design; behavioral targeting; dark patterns | Medium-High | Manipulative recommendation algorithms; deceptive chatbots |
| Dignity violations | AI applications that demean, objectify, or violate human dignity | Reputational damage; stakeholder opposition; regulatory action | Insufficient ethical review; misaligned incentives | Medium | Exploitative emotion recognition; invasive surveillance | ||
| Systemic | Concentration of power | AI capabilities that entrench market dominance or social inequalities | Antitrust scrutiny; social opposition; regulatory intervention | Platform economics; data advantages; network effects | Low | AI-enabled market manipulation; entrenched monopolies | |
| Democratic erosion | AI applications that undermine democratic processes or public discourse | Political backlash; regulatory response; social instability | Misinformation generation; micro-targeting; deepfakes | Medium | Election interference; synthetic media manipulation | ||
| Environmental | Computational footprint | Energy consumption and carbon emissions from AI training and operation | Sustainability goal conflicts; stakeholder criticism; regulatory requirements | Large model training; inefficient infrastructure; scaling practices | Low | Large language model training emissions; data center energy use |
3.2. Technical Risks
3.2.1. Algorithmic Bias and Discrimination
3.2.2. Opacity and Inexplicability
3.2.3. Security Vulnerabilities
3.2.4. Reliability and Robustness
3.3. Organizational Risks
3.3.1. Operational Disruption and Integration Challenges
3.3.2. Workforce and Capability Implications
3.3.3. Governance and Accountability Structures
3.4. Societal Risks
3.4.1. Erosion of Autonomy and Human Agency
3.4.2. Systemic and Societal Implications
3.4.3. Environmental Considerations
4. Organizational Dimensions of AI Risk
4.1. Sector-Specific Risk Profiles
4.1.1. Healthcare Sector
4.1.2. Financial Services Sector
| Sector | Primary AI Applications | Unique Sector Characteristics | Predominant Risk Types | Sector-Specific Risk Manifestations | Key Regulatory/Legal Considerations | Notable Incidents/Cases | Emerging Best Practices |
|---|---|---|---|---|---|---|---|
| Healthcare and Life Sciences | Clinical decision support; medical imaging analysis; drug discovery; patient monitoring; administrative automation; precision medicine | Life-safety criticality; extensive regulation; professional liability; patient vulnerability; data sensitivity | Accuracy and reliability; bias in health outcomes; privacy; explainability; liability allocation | Diagnostic errors affecting treatment; algorithmic bias in risk scores by race/ethnicity; unauthorized health inferences; inability to explain recommendations to clinicians | FDA regulation of AI/ML medical devices; HIPAA privacy requirements; medical malpractice liability; informed consent obligations; CE marking (EU) | Optum algorithm racial bias in care allocation (Obermeyer et al., 2019); IBM Watson oncology concerns; Epic sepsis model performance issues | Clinical validation requirements; diverse training data mandates; human-in-the-loop for critical decisions; post-market surveillance; algorithmic impact assessments |
| Financial Services | Credit scoring and lending decisions; fraud detection; algorithmic trading; customer service automation; anti-money laundering; insurance underwriting | Extensive regulatory oversight; systemic risk potential; consumer protection focus; discrimination concerns; high-frequency decision-making | Discrimination in lending; market manipulation; systemic instability; opacity in consumer decisions; security vulnerabilities | Discriminatory credit denials; flash crashes from algorithmic trading; unexplainable loan rejections; biased insurance pricing | Fair lending laws (ECOA, FHA); CFPB oversight; SEC trading regulations; GDPR right to explanation; state insurance regulations; Basel III considerations | Apple Card gender bias allegations; flash crash events; discriminatory auto lending settlements | Fair lending testing protocols; model risk management (SR 11-7); algorithmic auditing; adverse action explanation systems; stress testing for AI models |
| Criminal Justice and Public Safety | Recidivism risk assessment; facial recognition; predictive policing; evidence analysis; surveillance systems | Constitutional protections; due process requirements; civil liberties concerns; racial justice implications; high-stakes individual consequences | Bias perpetuating historical discrimination; due process violations; privacy and surveillance overreach; opacity in consequential decisions | Racially biased risk scores; wrongful identifications; over-policing of minority communities; inability to challenge algorithmic assessments | Constitutional due process protections; Fourth Amendment considerations; state facial recognition bans; CJIS requirements; consent decree requirements | COMPAS recidivism tool bias (Angwin et al., 2016); wrongful facial recognition arrests; predictive policing discrimination concerns | Independent algorithmic audits; mandatory human review; transparency requirements; moratoriums on high-risk applications; community oversight boards |
| Human Resources and Employment | Resume screening; candidate assessment; interview analysis; performance evaluation; workforce planning; employee monitoring | Employment discrimination laws; power asymmetries; worker privacy; collective bargaining implications | Hiring discrimination; privacy invasion; fairness in evaluations; worker surveillance concerns | Screening out protected groups; biased video interview analysis; invasive productivity monitoring; discriminatory performance ratings | Title VII and EEOC guidance; ADA considerations; GDPR employment provisions; state biometric laws; emerging AI hiring laws (NYC Local Law 144) | Amazon hiring tool gender bias; HireVue concerns; Illinois BIPA litigation | Adverse impact testing; third-party audits; candidate notification requirements; human review of automated rejections; transparency in evaluation criteria |
| Autonomous Systems and Transportation | Self-driving vehicles; aviation autopilot; drone operations; logistics optimization; traffic management | Physical safety primacy; complex liability allocation; infrastructure integration; public space operation | Safety-critical failures; liability uncertainty; cybersecurity vulnerabilities; ethical decision-making in emergencies | Collision fatalities; unclear crash liability; vehicle system hacking; trolley problem scenarios | NHTSA automated vehicle guidance; FAA drone regulations; state autonomous vehicle laws; product liability doctrine; international standards (ISO/SAE) | Tesla Autopilot fatalities; Uber autonomous vehicle pedestrian death; Boeing 737 MAX MCAS failures | Operational design domain specifications; disengagement reporting; safety case frameworks; graduated deployment; mandatory incident reporting |
| Education | Adaptive learning systems; automated grading; student performance prediction; administrative automation; proctoring systems | Vulnerable populations (minors); developmental considerations; equity concerns; educational mission alignment | Bias affecting educational opportunities; privacy of minors; surveillance concerns; equity in access | Biased tracking into educational pathways; student data commercialization; invasive exam proctoring; achievement gap amplification | FERPA privacy protections; COPPA for younger students; IEP requirements; state student privacy laws; civil rights obligations | Proctoring software bias and disability discrimination; student data breaches; adaptive learning equity concerns | Parental consent requirements; bias testing in educational algorithms; data minimization; human review of high-stakes decisions; equity impact assessments |
| Content Moderation and Media | Content recommendation; misinformation detection; content filtering; synthetic media detection; advertising targeting | Free expression considerations; platform scale; content velocity; cultural context variability | Censorship concerns; amplification of harmful content; political bias allegations; deepfake proliferation | Over-removal of legitimate speech; viral misinformation spread; algorithmic radicalization; synthetic media manipulation | Section 230 considerations; Digital Services Act (EU); election integrity laws; advertising disclosure requirements; right to be forgotten | Facebook algorithmic amplification concerns; YouTube radicalization studies; deepfake political manipulation | Transparency reports; appeal mechanisms; human review for complex cases; content provenance systems; researcher data access |
| Retail and Consumer Services | Personalized recommendations; dynamic pricing; inventory optimization; customer service chatbots; demand forecasting | Consumer protection focus; competitive dynamics; personalization expectations; price sensitivity | Price discrimination; manipulative personalization; consumer privacy; deceptive practices | Discriminatory pricing by demographics; addictive design patterns; excessive behavioral tracking; misleading chatbot interactions | FTC unfair and deceptive practices authority; state consumer protection laws; GDPR consent requirements; price discrimination concerns | Amazon pricing algorithm concerns; targeted advertising discrimination; dark pattern enforcement | Price transparency requirements; opt-out mechanisms; clear bot disclosure; algorithmic pricing audits; personalization controls |
4.1.3. Criminal Justice Sector
4.2. Organizational Maturity and Capabilities
4.3. Supply Chain and Vendor Considerations
5. Governance Approaches and Regulatory Frameworks
5.1. International and Multi-Stakeholder Initiatives
5.2. National and Regional Regulatory Approaches
5.3. Organizational Governance Structures
5.4. Challenges in AI Governance Implementation
6. Mitigation Strategies and Best Practices
6.1. Technical Interventions
6.1.1. Algorithmic Auditing and Testing
| Strategy Category | Specific Strategy | Target Risk Types | Implementation Level | Description and Key Activities | Resource Requirements | Implementation Complexity | Evidence of Effectiveness | Key Implementation Challenges | Enabling Standards/Frameworks |
|---|---|---|---|---|---|---|---|---|---|
| Technical Interventions | Algorithmic auditing | Bias; discrimination; accuracy; compliance | Technical; Organizational | Systematic examination of AI systems for bias, accuracy, and compliance through statistical testing, outcome analysis, and fairness metric evaluation | High (specialized expertise, tools, ongoing commitment) | High | Strong evidence for bias detection; emerging methods for other risks | Audit scope definition; benchmark selection; intersectional analysis complexity; remediation pathways | IEEE 7010; ISO/IEC 25010; NIST SP 1270; AIF360 toolkit |
| Explainable AI (XAI) implementation | Opacity; accountability gaps; trust deficits; regulatory compliance | Technical | Deploying interpretability and explainability techniques to make AI decision-making processes understandable to relevant stakeholders | High (technical expertise, computational resources, user interface design) | High | Growing evidence for enhanced trust and error detection; regulatory compliance benefits | Accuracy-explainability tradeoffs; stakeholder-appropriate explanations; local vs. global explanations | DARPA XAI program outcomes; ISO/IEC 22989; NIST AI RMF | |
| Robustness testing and adversarial training | Adversarial attacks; robustness failures; security vulnerabilities | Technical | Systematic testing of AI systems against adversarial inputs and edge cases, incorporating adversarial examples in training | Medium-High (security expertise, testing infrastructure) | Medium-High | Strong evidence for improved adversarial robustness; ongoing arms race dynamics | Comprehensive attack surface coverage; computational costs; novel attack vectors | NIST Adversarial ML taxonomy; MITRE ATLAS; CleverHans library | |
| Privacy-preserving techniques | Privacy violations; data protection compliance; data sensitivity | Technical | Implementing differential privacy, federated learning, homomorphic encryption, and other techniques to protect individual privacy | Medium-High (specialized technical expertise, potential performance tradeoffs) | High | Strong theoretical foundations; growing practical implementations | Utility-privacy tradeoffs; implementation complexity; performance overhead | NIST Privacy Framework; ISO/IEC 27701; GDPR technical measures | |
| Continuous monitoring and drift detection | Accuracy degradation; model drift; emerging bias; operational failures | Technical; Operational | Ongoing surveillance of AI system performance, data distributions, and outcome patterns to detect degradation or drift | Medium (monitoring infrastructure, alert systems, response protocols) | Medium | Strong evidence for early problem detection; essential for production systems | Alert fatigue; appropriate threshold setting; root cause analysis | MLOps practices; ISO/IEC 5338; Google ML best practices | |
| Data quality management | Data quality issues; bias from data; accuracy problems | Technical; Organizational | Systematic processes for ensuring training and operational data accuracy, completeness, representativeness, and currency | Medium (data governance infrastructure, quality tools, ongoing maintenance) | Medium | Strong evidence linking data quality to model performance and fairness | Legacy data challenges; representation assessment; data provenance tracking | DAMA-DMBOK; ISO 8000; FAIR principles | |
| Governance Mechanisms | AI ethics committees/review boards | Ethical risks; strategic risks; reputational risks; societal impacts | Governance | Establishing cross-functional bodies to review AI initiatives for ethical implications, provide guidance, and escalate concerns | Low-Medium (committee time, secretariat support) | Medium | Limited systematic evidence; growing adoption; variable effectiveness | Authority and influence; expertise composition; workflow integration; avoiding rubber-stamp dynamic | IEEE 7000 series; organizational ethics frameworks |
| Algorithmic impact assessments | All risk categories (comprehensive assessment) | Governance; Organizational | Structured pre-deployment and ongoing evaluations of AI system impacts on individuals, groups, and society | Medium (assessment frameworks, expertise, stakeholder engagement) | Medium-High | Emerging evidence from mandatory implementations; conceptual support from privacy impact assessment analogs | Standardization challenges; assessment quality variation; scope determination | Canada AIA framework; proposed EU requirements; AI Now Institute model | |
| Clear accountability structures | Accountability gaps; governance failures; oversight deficits | Governance; Organizational | Establishing explicit roles, responsibilities, and escalation paths for AI development, deployment, and oversight | Low-Medium (organizational design, policy development, role definition) | Medium | Theoretical support from corporate governance literature; limited AI-specific evidence | Distributed responsibility challenges; technical-business coordination; evolving systems | NIST AI RMF Govern function; RACI frameworks; ISO 38500 | |
| Third-party auditing and certification | Compliance risks; trust deficits; accountability gaps | Governance; External | Engaging independent external parties to assess AI system compliance, fairness, and trustworthiness | Medium-High (audit costs, preparation effort, remediation) | Medium | Growing evidence from financial model validation practices; emerging AI-specific evidence | Auditor expertise and independence; standard maturity; audit scope and depth | Emerging audit standards; SOC 2 for AI; proposed EU conformity assessment | |
| Policy and standards development | All risk categories (organizational baseline) | Governance | Creating organizational policies, standards, and guidelines governing AI development and use | Low-Medium (policy development expertise, stakeholder consultation) | Low-Medium | Foundation for other governance mechanisms; effectiveness depends on implementation | Policy enforcement; keeping pace with technology; practical applicability | ISO/IEC 42001 (AI management systems); internal policy frameworks | |
| Operational Practices | Human-in-the-loop processes | Automation failures; ethical issues; high-stakes decisions; edge cases | Operational | Designing AI systems to maintain meaningful human oversight, review, and intervention capabilities | Medium (workflow design, training, capacity allocation) | Medium | Strong evidence for error catching in high-stakes domains; concerns about automation bias | Automation bias mitigation; scalability; meaningful vs. superficial oversight | EU AI Act requirements; FDA guidance; aviation human factors standards |
| Red teaming and adversarial testing | Security vulnerabilities; robustness failures; unexpected behaviors; misuse potential | Operational; Technical | Dedicated teams attempting to find vulnerabilities, failure modes, and potential misuse in AI systems | Medium (specialized expertise, dedicated resources) | Medium | Growing evidence from cybersecurity applications; emerging AI-specific practices | Expertise availability; comprehensive scope; organizational receptivity to findings | Microsoft Responsible AI Standard; Anthropic practices; MITRE frameworks | |
| Incident response and learning systems | All operational risks; governance failures | Operational; Organizational | Establishing processes for detecting, responding to, escalating, and learning from AI-related incidents | Medium (response protocols, investigation capability, feedback loops) | Medium | Strong evidence from safety-critical industries; emerging AI-specific applications | Incident detection; root cause analysis for complex systems; organizational learning barriers | NIST Cybersecurity Framework; ISO 27035; safety management systems | |
| Staged deployment and rollback capabilities | Operational risks; unforeseen impacts; integration failures | Operational; Technical | Implementing gradual rollout processes with monitoring and ability to quickly reverse deployments | Medium (deployment infrastructure, monitoring, rollback mechanisms) | Medium | Strong evidence from software engineering practices; applicable to AI systems | Rollback complexity for learned systems; canary deployment design | DevOps/MLOps practices; site reliability engineering | |
| Stakeholder engagement and feedback | All risk categories; trust deficits; unforeseen impacts | Operational; Governance | Systematically engaging affected stakeholders in AI design, deployment, and ongoing governance | Low-Medium (engagement processes, feedback mechanisms, responsiveness) | Medium | Theoretical support from participatory design; emerging evidence from AI applications | Representative participation; meaningful influence; balancing diverse perspectives | Participatory design methods; community engagement frameworks | |
| Organizational Capabilities | AI literacy and training programs | Skill gaps; oversight failures; governance gaps; adoption risks | Organizational | Developing organization-wide understanding of AI capabilities, limitations, and responsible use | Medium (training development, delivery, ongoing updates) | Low-Medium | General evidence for training effectiveness; emerging AI-specific applications | Role-appropriate training design; keeping pace with technology; measuring effectiveness | Emerging AI literacy frameworks; professional development standards |
| Cross-functional AI governance teams | Governance gaps; siloed perspectives; coordination failures | Organizational | Establishing teams combining technical, legal, ethical, business, and domain expertise for AI governance | Medium (team formation, coordination mechanisms) | Medium | Theoretical support from interdisciplinary governance; limited systematic evidence | Expertise integration; decision-making authority; avoiding paralysis | Multidisciplinary team frameworks; matrix organization principles | |
| Vendor management and due diligence | Vendor dependence; third-party risks; supply chain vulnerabilities | Organizational | Rigorous assessment and ongoing oversight of external AI providers, tools, and components | Medium (due diligence processes, contract management, ongoing monitoring) | Medium | Evidence from IT vendor management; applicable to AI contexts | AI-specific assessment criteria; ongoing monitoring; contractual protections | ISO 27036; third-party risk management frameworks; AI-specific due diligence checklists | |
| Documentation and model cards | Opacity; accountability gaps; knowledge management; reproducibility | Organizational; Technical | Systematic documentation of AI system design, training, limitations, and intended use | Low-Medium (documentation standards, templates, maintenance processes) | Low | Growing evidence for improved understanding and appropriate use | Documentation maintenance; appropriate detail level; accessibility | Model cards (Mitchell et al., 2019); datasheets for datasets; system cards |
6.1.2. Explainability Implementation
6.1.3. Security Hardening
6.2. Governance Mechanisms
6.2.1. Algorithmic Impact Assessment
6.2.2. Ethics Review and Oversight
6.2.3. Documentation and Transparency
6.3. Operational Practices
6.3.1. Human Oversight and Control
6.3.2. Staged Deployment and Monitoring
6.3.3. Incident Response and Learning
6.4. Organizational Capabilities
6.4.1. Workforce Development
6.4.2. Cross-Functional Integration
7. Discussion and Future Directions
7.1. Implementation Challenges
7.2. Emerging Risk Areas
7.3. Toward Mature AI Risk Management
8. Conclusions
References
- Acemoglu, D.; Restrepo, P. Automation and new tasks: How technology displaces and reinstates labor. Journal of Economic Perspectives 2019, 33(2), 3–30. [Google Scholar] [CrossRef]
- Amershi, S.; Begel, A.; Bird, C.; DeLine, R.; Gall, H.; Kamar, E.; Nagappan, N.; Nushi, B.; Zimmermann, T. Software engineering for machine learning: A case study. In Proceedings of the 41st International Conference on Software Engineering: Software Engineering in Practice; 2019; pp. 291–300. [Google Scholar]
- Amodei, D.; Olah, C.; Steinhardt, J.; Christiano, P.; Schulman, J.; Mané, D. Concrete problems in AI safety. arXiv 2016, arXiv:1606.06565. [Google Scholar] [CrossRef]
- Angwin, J.; Larson, J.; Mattu, S.; Kirchner, L. Machine bias. In ProPublica; 23 May 2016. [Google Scholar]
- Arrieta, A. B.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; García, S.; Gil-López, S.; Molina, D.; Benjamins, R.; Chatila, R.; Herrera, F. Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Information Fusion 58 2020, 82–115. [Google Scholar] [CrossRef]
- Barocas, S.; Selbst, A. D. Big data's disparate impact. California Law Review 2016, 104(3), 671–732. [Google Scholar] [CrossRef]
- Board of Governors of the Federal Reserve System. Supervisory guidance on model risk management (SR Letter 11-7); Federal Reserve System, 2011. [Google Scholar]
- Bostrom, N. Superintelligence: Paths, dangers, strategies; Oxford University Press, 2014. [Google Scholar]
- Bradford, A. Digital empires: The global battle to regulate technology; Oxford University Press, 2023. [Google Scholar]
- Breck, E.; Cai, S.; Nielsen, E.; Salib, M.; Sculley, D. The ML test score: A rubric for ML production readiness and technical debt reduction. In Proceedings of the 2017 IEEE International Conference on Big Data; 2017; pp. 1123–1132. [Google Scholar]
- Buchanan, B. G. Artificial intelligence in finance. In The Alan Turing Institute; 2019. [Google Scholar]
- Buolamwini, J.; Gebru, T. Gender shades: Intersectional accuracy disparities in commercial gender classification. In Proceedings of the 1st Conference on Fairness, Accountability and Transparency; 2018; pp. 77–91. [Google Scholar]
- Burrell, J. How the machine "thinks": Understanding opacity in machine learning algorithms. Big Data & Society 2016, 3(1), 1–12. [Google Scholar] [CrossRef]
- Carr, N. The glass cage: Automation and us; W. W. Norton & Company, 2014. [Google Scholar]
- Cath, C.; Wachter, S.; Mittelstadt, B.; Taddeo, M.; Floridi, L. Artificial intelligence and the "good society": The US, EU, and UK approach. Science and Engineering Ethics 2018, 24(2), 505–528. [Google Scholar]
- Chouldechova, A. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big Data 2017, 5(2), 153–163. [Google Scholar] [CrossRef]
- Crawford, K. Atlas of AI: Power, politics, and the planetary costs of artificial intelligence; Yale University Press, 2021. [Google Scholar]
- Dastin, J. Amazon scraps secret AI recruiting tool that showed bias against women. In Reuters; 10 October 2018. [Google Scholar]
- Davenport, T. H.; Ronanki, R. Artificial intelligence for the real world. Harvard Business Review 2018, 96(1), 108–116. [Google Scholar]
- Doshi-Velez, F.; Kim, B. Towards a rigorous science of interpretable machine learning. arXiv 2017, arXiv:1702.08608. [Google Scholar] [CrossRef]
- Edmondson, A. C. The fearless organization: Creating psychological safety in the workplace for learning, innovation, and growth; Wiley, 2019. [Google Scholar]
- European Commission. Regulation (EU) 2024/1689 of the European Parliament and of the Council laying down harmonised rules on artificial intelligence (Artificial Intelligence Act). Official Journal of the European Union 2024. [Google Scholar]
- FDA. Artificial intelligence/machine learning (AI/ML)-based software as a medical device (SaMD) action plan; U.S. Food and Drug Administration, 2021. [Google Scholar]
- Floridi, L.; Cowls, J.; Beltrametti, M.; Chatila, R.; Chazerand, P.; Dignum, V.; Luetge, C.; Madelin, R.; Pagallo, U.; Rossi, F.; Schafer, B.; Valcke, P.; Vayena, E. AI4People—An ethical framework for a good AI society: Opportunities, risks, principles, and recommendations. Minds and Machines 2018, 28(4), 689–707. [Google Scholar] [CrossRef]
- Gama, J.; Žliobaitė, I.; Bifet, A.; Pechenizkiy, M.; Bouchachia, A. A survey on concept drift adaptation. ACM Computing Surveys 2014, 46(4), 1–37. [Google Scholar] [CrossRef]
- Gartner. Gartner survey reveals leading organizations expect to double the number of AI projects within the next year; Gartner, 2019. [Google Scholar]
- Gebru, T.; Morgenstern, J.; Vecchione, B.; Vaughan, J. W.; Wallach, H.; Daumé, H., III; Crawford, K. Datasheets for datasets. Communications of the ACM 2021, 64(12), 86–92. [Google Scholar] [CrossRef]
- Goodfellow, I. J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. In Proceedings of the 3rd International Conference on Learning Representations; 2015. [Google Scholar]
- Government of Canada. Algorithmic Impact Assessment Tool; Treasury Board of Canada Secretariat, 2022. [Google Scholar]
- Green, B.; Chen, Y. The principles and limits of algorithm-in-the-loop decision making. Proceedings of the ACM on Human-Computer Interaction 2019, 3(CSCW), 1–24. [Google Scholar] [CrossRef]
- Gu, T.; Dolan-Gavitt, B.; Garg, S. BadNets: Identifying vulnerabilities in the machine learning model supply chain. arXiv 2017, arXiv:1708.06733. [Google Scholar]
- ISO. ISO 31000:2018 Risk management—Guidelines. International Organization for Standardization, 2018.
- Jordan, M. I.; Mitchell, T. M. Machine learning: Trends, perspectives, and prospects. Science 2015, 349(6245), 255–260. [Google Scholar] [CrossRef]
- Kellogg, K. C.; Valentine, M. A.; Christin, A. Algorithms at work: The new contested terrain of control. Academy of Management Annals 2020, 14(1), 366–410. [Google Scholar] [CrossRef]
- Kessler, S.; Martin, K.; Ransbotham, S.; Kiron, D. The cultural benefits of artificial intelligence in the enterprise; MIT Sloan Management Review; 2022. [Google Scholar]
- Leveson, N. Engineering a safer world: Systems thinking applied to safety; MIT Press, 2011. [Google Scholar]
- Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards deep learning models resistant to adversarial attacks. In Proceedings of the 6th International Conference on Learning Representations; 2018. [Google Scholar]
- Marchetti, R. AI governance: Challenges, opportunities, and recommendations. Global Policy 2021, 12(S6), 43–49. [Google Scholar]
- Metcalf, J.; Moss, E.; Watkins, E. A.; Singh, R.; Elish, M. C. Algorithmic impact assessments and accountability: The co-construction of impacts. In Proceedings of the 2019 Conference on Fairness, Accountability, and Transparency; 2019; pp. 735–746. [Google Scholar]
- Metcalf, J.; Moss, E.; boyd, d. Owning ethics: Corporate logics, Silicon Valley, and the institutionalization of ethics. Social Research 2019, 86(2), 449–476. [Google Scholar] [CrossRef]
- Miller, T. Explanation in artificial intelligence: Insights from the social sciences. Artificial Intelligence 267 2019, 1–38. [Google Scholar] [CrossRef]
- Mitchell, M.; Wu, S.; Zaldivar, A.; Barnes, P.; Vasserman, L.; Hutchinson, B.; Spitzer, E.; Raji, I. D.; Gebru, T. Model cards for model reporting. In Proceedings of the Conference on Fairness, Accountability, and Transparency; 2019; pp. 220–229. [Google Scholar]
- Mitchell, S.; Potash, E.; Barocas, S.; D'Amour, A.; Lum, K. Algorithmic fairness: Choices, assumptions, and definitions. Annual Review of Statistics and Its Application 8 2021, 141–163. [Google Scholar] [CrossRef]
- Mittelstadt, B. D.; Allo, P.; Taddeo, M.; Wachter, S.; Floridi, L. The ethics of algorithms: Mapping the debate. Big Data & Society 2016, 3(2), 1–21. [Google Scholar] [CrossRef]
- NIST. Framework for improving critical infrastructure cybersecurity (Version 1.1); National Institute of Standards and Technology, 2018. [Google Scholar]
- NIST. Artificial intelligence risk management framework (AI RMF 1.0); National Institute of Standards and Technology, 2023. [Google Scholar]
- Nissenbaum, H. Accountability in a computerized society. Science and Engineering Ethics 1996, 2(1), 25–42. [Google Scholar] [CrossRef]
- Obermeyer, Z.; Powers, B.; Vogeli, C.; Mullainathan, S. Dissecting racial bias in an algorithm used to manage the health of populations. Science 2019, 366(6464), 447–453. [Google Scholar] [CrossRef] [PubMed]
- OECD. Recommendation of the Council on Artificial Intelligence; OECD/LEGAL/0449. Organisation for Economic Co-operation and Development, 2019. [Google Scholar]
- O'Neil, C. Weapons of math destruction: How big data increases inequality and threatens democracy; Crown, 2016. [Google Scholar]
- Raji, I. D.; Smart, A.; White, R. N.; Mitchell, M.; Gebru, T.; Hutchinson, B.; Smith-Loud, J.; Theron, D.; Barnes, P. Closing the AI accountability gap: Defining an end-to-end framework for internal algorithmic auditing. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency; 2020; pp. 33–44. [Google Scholar]
- Ransbotham, S.; Kiron, D.; Gerbert, P.; Reeves, M. Reshaping business with artificial intelligence. MIT Sloan Management Review 2017, 59(1), 1–17. [Google Scholar]
- Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence 2019, 1(5), 206–215. [Google Scholar] [CrossRef]
- Russell, S.; Norvig, P. Artificial intelligence: A modern approach, 4th ed.; Pearson, 2021. [Google Scholar]
- Scherer, M. U. Regulating artificial intelligence systems: Risks, challenges, competencies, and strategies. Harvard Journal of Law & Technology 2016, 29(2), 353–400. [Google Scholar]
- Selbst, A. D. An institutional view of algorithmic impact assessments. Harvard Journal of Law & Technology 2021, 35(1), 117–191. [Google Scholar]
- Selbst, A. D.; boyd, d.; Friedler, S. A.; Venkatasubramanian, S.; Vertesi, J. Fairness and abstraction in sociotechnical systems. In Proceedings of the Conference on Fairness, Accountability, and Transparency; 2019; pp. 59–68. [Google Scholar]
- Selbst, A. D.; Powles, J. Meaningful information and the right to explanation. International Data Privacy Law 2017, 7(4), 233–242. [Google Scholar] [CrossRef]
- Skitka, L. J.; Mosier, K. L.; Burdick, M. Does automation bias decision-making? International Journal of Human-Computer Studies 1999, 51(5), 991–1006. [Google Scholar] [CrossRef]
- Sloane, M.; Moss, E.; Awomolo, O.; Forlano, L. Participation is not a design fix for machine learning. In Proceedings of the 37th International Conference on Machine Learning Workshop on Participatory Approaches to Machine Learning; 2020. [Google Scholar]
- Smuha, N. A. From a "race to AI" to a "race to AI regulation": Regulatory competition for artificial intelligence. Law, Innovation and Technology 2021, 13(1), 57–84. [Google Scholar] [CrossRef]
- Strubell, E.; Ganesh, A.; McCallum, A. Energy and policy considerations for deep learning in NLP. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics; 2019; pp. 3645–3650. [Google Scholar]
- Topol, E. J. High-performance medicine: The convergence of human and artificial intelligence. Nature Medicine 2019, 25(1), 44–56. [Google Scholar] [CrossRef] [PubMed]
- Tramèr, F.; Zhang, F.; Juels, A.; Reiter, M. K.; Ristenpart, T. Stealing machine learning models via prediction APIs. In Proceedings of the 25th USENIX Security Symposium; 2016; pp. 601–618. [Google Scholar]
- UNESCO. Recommendation on the ethics of artificial intelligence; United Nations Educational, Scientific and Cultural Organization, 2021. [Google Scholar]
- Weidinger, L.; Mellor, J.; Rauh, M.; Griffin, C.; Uesato, J.; Huang, P. S.; Cheng, M.; Glaese, M.; Balle, B.; Kasirzadeh, A.; Kenton, Z.; Brown, S.; Hawkins, W.; Stepleton, T.; Biles, C.; Birhane, A.; Haas, J.; Rimell, L.; Hendricks, L. A.; Gabriel, I. Taxonomy of risks posed by language models. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency; 2022; pp. 214–229. [Google Scholar]
- House, White. Blueprint for an AI Bill of Rights: Making automated systems work for the American people. In Office of Science and Technology Policy; 2022. [Google Scholar]
- House, White. Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence. Executive Order 14110; 2023.
- Yeung, K. Hypernudge": Big Data as a mode of regulation by design. Information, Communication & Society 2017, 20(1), 118–136. [Google Scholar]
- Zuboff, S. The age of surveillance capitalism: The fight for a human future at the new frontier of power; PublicAffairs, 2019. [Google Scholar]


| Framework | Region/Scope | Year | Core Principles | Legal Status | Risk Classification Approach | Enforcement Mechanisms | Key Strengths | Notable Limitations |
|---|---|---|---|---|---|---|---|---|
| EU Artificial Intelligence Act | European Union (27 member states) | 2024 | Human oversight, transparency, accountability, non-discrimination, safety, data governance | Binding regulation | Four-tier system: Unacceptable, High, Limited, Minimal risk | Fines up to €35 million or 7% global turnover; national supervisory authorities; EU AI Office | Comprehensive scope; legally binding; harmonized standards; clear prohibitions | Implementation complexity; potential innovation barriers; extraterritorial challenges |
| NIST AI Risk Management Framework | United States | 2023 | Trustworthy AI characteristics: valid, reliable, safe, secure, resilient, accountable, transparent, explainable, privacy-enhanced, fair | Voluntary guidance | Context-dependent; organization-specific assessment | Non-binding; voluntary adoption; no direct penalties | Flexible; sector-adaptable; strong technical guidance; stakeholder input | Lacks enforcement power; inconsistent adoption; limited accountability mechanisms |
| Singapore Model AI Governance Framework | Singapore | 2020 (2nd ed.) | Human-centric AI; explainability; transparency; fairness; human oversight | Voluntary framework | Sector-specific; probability and severity matrix | Industry self-regulation; sectoral guidelines | Practical implementation focus; business-friendly; clear guidance | Limited to voluntary adoption; small jurisdiction scope |
| OECD AI Principles | International (38 member countries + partners) | 2019 | Inclusive growth; sustainable development; human-centered values; transparency; robustness; accountability | Soft law/Recommendation | General principles; context-sensitive | Peer review; national implementation monitoring | Broad international consensus; foundational influence; multi-stakeholder approach | Non-binding; variable national implementation; lacks specificity |
| China New Generation AI Governance Principles | China | 2019/2021 | Harmony, fairness, inclusivity, respect for privacy, safety, shared responsibility | State-guided principles with emerging regulations | Sector-specific regulations emerging | State oversight; algorithmic registry requirements; sectoral enforcement | Large-scale implementation; rapid regulatory development | Limited transparency; state-centric approach; human rights concerns |
| Canada Directive on Automated Decision-Making | Canada (Federal government) | 2019 | Transparency; accountability; legality; procedural fairness | Binding for federal agencies | Four-level impact assessment system | Treasury Board oversight; mandatory compliance for federal bodies | Clear public sector guidance; impact assessment model; transparency requirements | Limited to federal government; private sector gap |
| UK AI Regulation White Paper (Pro-Innovation Approach) | United Kingdom | 2023 | Safety; transparency; fairness; accountability; contestability | Principles-based; sector-specific regulation | Context-dependent; regulator-led assessment | Existing sectoral regulators; no central AI authority | Flexible; innovation-friendly; leverages existing expertise | Regulatory fragmentation risk; potential gaps; coordination challenges |
| Brazil AI Bill (PL 2338/2023) | Brazil | 2023 (pending) | Human dignity; non-discrimination; transparency; accountability; security | Proposed binding legislation | Risk-based tiering similar to EU approach | National AI authority proposed; administrative penalties | Comprehensive scope; rights-based approach | Still under legislative consideration; implementation uncertain |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
