1. Introduction
The global health landscape is experiencing an unprecedented digital transformation, with artificial intelligence (AI) and machine learning (ML) technologies increasingly integrated into healthcare delivery systems worldwide [
1]. This technological revolution is particularly pronounced in low- and middle-income countries (LMICs), where digital health innovations offer opportunities to leapfrog traditional healthcare infrastructure limitations [
2]. Mobile health platforms, AI-powered diagnostic tools, and algorithmic clinical decision support systems are being rapidly deployed across diverse LMIC contexts.
However, this rapid technological adoption occurs within complex socioeconomic environments characterized by significant health inequities, limited regulatory frameworks, and varying degrees of digital literacy [
3]. Unlike high-income countries where AI implementation typically occurs within well-established healthcare systems with robust oversight mechanisms, LMICs often lack the institutional capacity to adequately evaluate, monitor, and mitigate potential algorithmic biases [
4].
Algorithmic bias in healthcare encompasses systematic and unfair discrimination embedded within computational systems that affects clinical decision-making, resource allocation, and health service delivery [
5]. These biases can manifest through various mechanisms, including biased training data, inappropriate algorithmic design choices, and inadequate consideration of population diversity during system development and deployment [
4].
The intersection between algorithmic bias and social determinants of health represents a critical area of investigation that remains theoretically underdeveloped and empirically understudied [
6]. Social determinants of health encompass the conditions in which people are born, grow, live, work, and age, including the broader set of forces and systems that shape daily life conditions [
7]. Algorithmic bias in healthcare may function as a novel mechanism through which traditional social determinants of health operate in digitalized healthcare systems [
8].
Current evidence regarding algorithmic bias in healthcare derives predominantly from high-income country contexts, with limited empirical data available from LMIC settings [
9]. Existing research methodologies may be inadequately suited for LMIC contexts, where data availability, regulatory frameworks, and healthcare delivery models differ substantially [
10]. Investigating algorithmic bias in LMIC healthcare systems presents unique methodological challenges requiring specialized research approaches.
The primary aim of this work is to develop a comprehensive, replicable methodological framework for systematically investigating algorithmic bias as a mechanism affecting health equity in LMIC healthcare systems. This framework will provide standardized approaches for identifying, measuring, and analyzing algorithmic bias while maintaining sensitivity to diverse cultural, economic, and healthcare system contexts.
2. Materials and Methods
2.1. Overall Study Design
This research protocol employs a sequential explanatory mixed-methods design implemented across three interconnected phases over 36 months. The methodology combines systematic evidence synthesis, primary data collection, and framework development to generate comprehensive understanding of algorithmic bias manifestation and impact in LMIC healthcare contexts.
The research design incorporates a multi-country comparative framework encompassing diverse LMIC contexts representing different geographical regions, economic development levels, and healthcare system structures. Quality assurance mechanisms include regular methodology review meetings, external expert consultation, and systematic documentation of methodological decisions.
2.2. Theoretical Framework
The social determinants of health framework, as conceptualized by the World Health Organization, provides the theoretical foundation for understanding how algorithmic bias may function as a mechanism affecting health equity [
7]. This framework identifies structural determinants including socioeconomic stratification, governance systems, and cultural norms as fundamental drivers of health inequities.
This research protocol proposes a novel theoretical framework conceptualizing algorithmic bias as an emergent social determinant of health that operates through direct and indirect pathways to influence health outcomes in LMIC contexts. The conceptual model identifies three primary pathways: (1) direct discrimination through systematic differential treatment recommendations; (2) indirect discrimination through perpetuation of existing healthcare disparities; and (3) systemic exclusion through underrepresentation in algorithmic system development processes.
2.3. Phase 1: Systematic Evidence Mapping
2.3.1. Scoping Review Protocol
The systematic evidence mapping begins with a comprehensive scoping review following the PRISMA-ScR framework [
11]. The literature search strategy encompasses multiple electronic databases including PubMed, Embase, Scopus, IEEE Xplore, and Web of Science, plus regional databases and grey literature sources.
Search terms combine concepts related to algorithmic bias, healthcare applications, and LMIC contexts. The temporal scope covers publications from 2015 to present. Inclusion criteria encompass peer-reviewed articles, conference papers, and reports addressing algorithmic bias in healthcare AI applications within LMIC settings.
Data extraction utilizes standardized forms capturing study characteristics, AI application details, bias assessment methods, findings, and policy implications. Two independent reviewers conduct screening and data extraction with disagreements resolved through consensus discussion.
2.3.2. Policy Landscape Analysis
The policy landscape analysis provides systematic examination of existing AI governance frameworks across selected LMIC contexts. Country selection employs purposive sampling ensuring representation across geographical regions, economic development levels, and healthcare system types across 15 LMICs.
Data sources encompass government websites, ministry publications, regulatory documents, national AI strategies, and reports from international organizations. The analytical framework employs systematic content analysis using standardized coding schemes addressing policy scope, governance mechanisms, and implementation provisions.
2.4. Phase 2: Primary Data Collection
2.4.1. Multi-Stakeholder Interview Protocol
Primary data collection employs semi-structured interviews with diverse stakeholders across multiple LMIC contexts. Participant recruitment employs purposive sampling with maximum variation ensuring representation across stakeholder groups, geographical locations, and organizational types.
Target participant categories include health system administrators and policymakers (n=30), healthcare providers (n=50), AI researchers and health informatics specialists (n=25), and patient advocacy representatives (n=20) across five countries: Nigeria, India, Kenya, Brazil, and Bangladesh.
Data collection employs secure video conferencing with provisions for telephone interviews where connectivity is limited. Interview guides address stakeholder experiences with AI in healthcare, perceptions of algorithmic bias, organizational practices for AI evaluation, and policy recommendations. Thematic analysis employs the framework approach with both deductive and inductive themes [
12].
2.4.2. Healthcare System Case Studies
In-depth case studies provide detailed examination of algorithmic bias manifestation within specific healthcare system contexts. Case study methodology follows established principles for explanatory case study research [
13]. Each case study employs multiple data sources including administrative health data, policy documentation, stakeholder interviews, and technical system documentation.
Analytical procedures employ cross-case pattern matching and explanation building to identify common mechanisms and context-specific factors [
14]. The selected countries represent diverse healthcare system characteristics and varying levels of AI implementation in healthcare delivery.
2.5. Phase 3: Framework Development and Validation
2.5.1. Integrated Analytical Framework
Framework development synthesizes evidence from all previous phases using established methodological approaches including framework synthesis and meta-ethnography [
15]. The synthesis process begins with systematic extraction of key findings using standardized templates.
Framework construction involves iterative development of conceptual models integrating empirical findings with existing theoretical frameworks. The resulting analytical framework encompasses theoretical models, operational definitions, methodological guidelines, and implementation strategies for different healthcare system contexts.
2.5.2. Protocol Validation and Refinement
The final phase involves validation and refinement through pilot implementation in Ghana and Vietnam. Pilot implementation focuses on protocol feasibility, cultural appropriateness, and methodological robustness.
Validation employs established criteria including reliability assessment, validity evaluation, and practicality assessment. Refinement processes incorporate stakeholder feedback and pilot findings to optimize protocol components for broader implementation.
2.6. Ethical Considerations
The ethical framework prioritizes justice, beneficence, respect for persons, and cultural sensitivity while addressing specific challenges associated with algorithmic bias research in LMIC contexts [
16]. Justice considerations encompass fair participant selection, equitable benefit distribution, and protection of vulnerable populations.
Institutional Review Board approval strategies involve comprehensive multi-country ethics review processes. Data protection measures comply with international standards including relevant GDPR provisions and national data protection laws. All participants will provide informed consent, and data will be anonymized to protect participant confidentiality.
2.7. Data Analysis
Quantitative data analysis will employ descriptive statistics and comparative analysis across countries and stakeholder groups. Qualitative data analysis will utilize thematic analysis following established frameworks [
12]. Mixed-methods integration will employ joint displays and meta-inferences to synthesize findings across data types.
Cross-case analysis will identify patterns and variations across different LMIC contexts. Framework development will employ systematic synthesis techniques to integrate findings into coherent theoretical and methodological frameworks.
3. Results
3.1. Expected Methodological Contributions
This research protocol will generate several significant methodological contributions including a comprehensive, validated research protocol specifically designed for LMIC contexts, standardized measurement tools and indicators, best practices for cross-cultural research implementation, and frameworks for policy-relevant research design.
The systematic evidence mapping will provide the first comprehensive assessment of existing knowledge regarding algorithmic bias in LMIC healthcare contexts. This will identify critical evidence gaps and inform priority areas for future research investment.
3.2. Anticipated Empirical Findings
Multi-stakeholder interviews will generate understanding of diverse perspectives and experiences regarding algorithmic bias in healthcare delivery. This will include healthcare provider experiences, patient advocacy concerns, policy maker priorities, and researcher perspectives on methodological challenges.
Healthcare system case studies will provide detailed evidence regarding specific mechanisms through which algorithmic bias affects health outcomes. Comparative analysis will identify common patterns and contextual variations enabling generalizable findings while recognizing important contextual factors.
3.3. Framework Development Outcomes
The integrated analytical framework will provide theoretical models linking algorithmic bias to health equity outcomes in LMIC contexts. This will include operational definitions suitable for empirical research, measurement indicators adaptable across diverse healthcare systems, and implementation guidelines accounting for resource constraints.
Validation processes will demonstrate framework reliability, validity, and practical applicability across different LMIC contexts. Refinement activities will optimize framework components based on pilot implementation experiences and stakeholder feedback.
4. Discussion
This research protocol addresses critical methodological gaps in algorithmic bias research within LMIC healthcare contexts. The multi-phase, mixed-methods approach provides balanced consideration of methodological rigor and practical implementation feasibility while maintaining sensitivity to complex cultural, economic, and healthcare system factors.
The significance extends beyond immediate research contributions to encompass broader implications for global health equity, international development, and technology governance. As AI technologies become increasingly integrated into healthcare delivery systems worldwide, ensuring equitable access while protecting vulnerable populations from algorithmic discrimination represents a fundamental challenge for global health and social justice.
4.1. Methodological Innovations
The proposed framework introduces several methodological innovations specifically designed for LMIC research contexts. The integration of systematic evidence mapping with primary data collection provides comprehensive understanding combining existing knowledge with novel empirical insights.
The multi-stakeholder approach ensures diverse perspectives are incorporated throughout the research process. This participatory methodology recognizes that algorithmic bias affects multiple stakeholder groups and requires inclusive investigation approaches.
4.2. Implementation Considerations
Successful implementation requires substantial investment in local research capacity and institutional partnerships [
17]. The capacity building strategy recognizes that sustainable research implementation requires long-term collaborative relationships with academic institutions, government agencies, and healthcare systems.
Resource requirements for this 36-month research protocol total approximately $2.5 million, including personnel costs, travel expenses, data collection costs, and capacity building activities. Cost-effectiveness optimization strategies maximize research impact through collaborative approaches and resource sharing agreements.
4.3. Policy Implications
The research will generate evidence-based recommendations for AI governance frameworks tailored to LMIC healthcare contexts, implementation guidelines for bias mitigation strategies, capacity building frameworks, and international cooperation strategies facilitating knowledge sharing and collaborative development.
These policy contributions will support evidence-based decision making regarding AI implementation in healthcare while promoting equitable access and protection of vulnerable populations.
4.4. Limitations
This research protocol acknowledges several important limitations including resource constraints, data availability variations, cultural and linguistic barriers, rapidly evolving technology landscape, and regulatory approval challenges that may affect implementation feasibility and findings generalizability.
The focus on selected LMIC contexts may limit generalizability to other settings. However, the multi-country comparative approach and framework validation processes are designed to maximize transferability while recognizing contextual variations.
5. Conclusions
This research protocol presents the first comprehensive methodological framework specifically designed for investigating algorithmic bias as a social determinant of health in LMIC healthcare contexts. By addressing critical evidence gaps and methodological limitations, this protocol establishes the foundation for systematic, rigorous, and culturally appropriate investigation across diverse global health settings.
Through systematic evidence mapping, comprehensive stakeholder engagement, detailed case study analysis, and collaborative framework development, this protocol generates the methodological tools and empirical evidence necessary for evidence-based policy development.
Implementation requires commitment from multiple stakeholder communities including academic institutions, funding organizations, policy makers, healthcare leaders, and civil society organizations. The collaborative approach reflects recognition that addressing algorithmic bias requires sustained international cooperation, knowledge sharing, and capacity building initiatives.
This protocol represents a call for international collaboration and coordinated action to ensure that the digital health transformation occurring across LMICs promotes rather than undermines health equity and social justice. Through systematic investigation, evidence-based policy development, and collaborative capacity building, the global health community can work together to harness AI benefits while protecting vulnerable populations from algorithmic discrimination.
Supplementary Materials
The following supporting information can be downloaded at: [URL], Interview guides; Policy analysis framework; Data extraction forms.
Informed Consent Statement
Informed consent was obtained from all subjects involved in the study.
Data Availability Statement
Conflicts of Interest
The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.
Abbreviations
The following abbreviations are used in this manuscript:
| AI |
Artificial Intelligence |
| ML |
Machine Learning |
| LMIC |
Low- and Middle-Income Countries |
| WHO |
World Health Organization |
| PRISMA-ScR |
Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews |
References
- Schwalbe, N.; Wahl, B. Artificial intelligence and the future of global health. Lancet 2020, 395, 1579–1586. [Google Scholar] [CrossRef] [PubMed]
- Alami, H.; Gagnon, M.P.; Fortin, J.P. Digital health and the challenge of health systems transformation. mHealth 2020, 6, 7. [Google Scholar] [CrossRef] [PubMed]
- Gwagwa, A.; Kraemer-Mbula, E.; Rizk, N.; Rutenberg, I.; De Beer, J. Artificial intelligence (AI) governance in Africa: A participatory approach. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency; ACM: New York, NY, USA, 2021; pp. 329–342. [Google Scholar]
- Rajkomar, A.; Hardt, M.; Howell, M.D.; Corrado, G.; Chin, M.H. Ensuring fairness in machine learning to advance health equity. Ann. Intern. Med. 2018, 169, 866–872. [Google Scholar] [CrossRef] [PubMed]
- Obermeyer, Z.; Powers, B.; Vogeli, C.; Mullainathan, S. Dissecting racial bias in an algorithm used to manage the health of populations. Science 2019, 366, 447–453. [Google Scholar] [CrossRef] [PubMed]
- Veinot, T.C.; Mitchell, H.; Ancker, J.S. Good intentions are not enough: How informatics interventions can worsen inequality. J. Am. Med. Inform. Assoc. 2018, 25, 1080–1088. [Google Scholar] [CrossRef] [PubMed]
- Marmot, M.; Wilkinson, R.G. Social Determinants of Health, 2nd ed.; Oxford University Press: Oxford, UK, 2006. [Google Scholar]
- Benjamin, R. Race After Technology: Abolitionist Tools for the New Jim Code; Polity Press: Cambridge, UK, 2019. [Google Scholar]
- Chen, I.Y.; Pierson, E.; Rose, S.; Joshi, S.; Ferryman, K.; Ghassemi, M. Ethical machine learning in healthcare. Annu. Rev. Biomed. Data Sci. 2021, 4, 123–144. [Google Scholar] [CrossRef] [PubMed]
- Reddy, S.; Allan, S.; Coghlan, S.; Cooper, P. A governance model for the application of AI in health care. J. Am. Med. Inform. Assoc. 2020, 27, 491–497. [Google Scholar] [CrossRef] [PubMed]
- Tricco, A.C.; Lillie, E.; Zarin, W.; O’Brien, K.K.; Colquhoun, H.; Levac, D.; Moher, D.; Peters, M.D.J.; Horsley, T.; Weeks, L.; et al. PRISMA extension for scoping reviews (PRISMA-ScR): Checklist and explanation. Ann. Intern. Med. 2018, 169, 467–473. [Google Scholar] [CrossRef] [PubMed]
- Gale, N.K.; Heath, G.; Cameron, E.; Rashid, S.; Redwood, S. Using the framework method for the analysis of qualitative data in multi-disciplinary health research. BMC Med. Res. Methodol. 2013, 13, 117. [Google Scholar] [CrossRef] [PubMed]
- Yin, R.K. Case Study Research and Applications: Design and Methods, 6th ed.; Sage Publications: Thousand Oaks, CA, USA, 2018. [Google Scholar]
- Miles, M.B.; Huberman, A.M.; Saldaña, J. Qualitative Data Analysis: A Methods Sourcebook, 3rd ed.; Sage Publications: Thousand Oaks, CA, USA, 2014. [Google Scholar]
- Carroll, C.; Booth, A.; Cooper, K. A worked example of “best fit” framework synthesis: A systematic review of views concerning the taking of some potential chemopreventive agents. BMC Med. Res. Methodol. 2013, 13, 117. [Google Scholar] [CrossRef] [PubMed]
- Emanuel, E.J.; Wendler, D.; Grady, C. What makes clinical research ethical? JAMA 2000, 283, 2701–2711. [Google Scholar] [CrossRef] [PubMed]
- Whitworth, J.; Sewankambo, N.K.; Snewin, V.A. Improving implementation: Building research capacity in maternal, neonatal, and child health in Africa. PLoS Med. 2010, 7, e1000299. [Google Scholar] [CrossRef] [PubMed]
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).