Submitted:
27 June 2024
Posted:
28 June 2024
You are already at the latest version
Abstract
Keywords:
1. Introduction
2. Literature Review
3. Methodology
3.1. Database Creation (Test and Training Data)
3.2. AI Model Selection and Comparative Analysis Across AI Models
3.3. Contextual and Intent-Based Design
3.4. AI Response Analysis and Bias Assessment
3.5. AI-BiasAudit Tool and Racial Data for Social Science Researchers and AI Developers
4. Results and Analysis
5. Discussion and Conclusion
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Baer, T. 2013. Process big data at speed. ComputerWeekly. Available at: https://www.computerweekly.com/feature/Process-big-data-at-speed.
- Baum, Jeremy, and John Villasenor. 2023. The politics of AI: ChatGPT and political bias. The Brookings Institution. Available at: https://www.brookings.edu/articles/the-politics-of-ai-chatgpt-and-political-bias/.
- Barocas, S., and A. D. Selbst. 2016. Big data's disparate impact. California Law Review 104(3): 671-732. Available at: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2477899.
- Bietti, E. 2023. A Genealogy of Digital Platform Regulation. Georgetown Law Technology Review 7(1): 1.
- Brown, S. 2021. Machine learning, explained. Ideas Made to Mater Artificial Intelligence. MIT Sloan School of Management. Available at: https://mitsloan.mit.edu/ideas-made-to-matter/machine-learning-explained.
- Brown, Tom B., et al. 2020. Language Models are Few-Shot Learners. Advances in Neural Information Processing Systems. Available at: https://dl.acm.org/doi/pdf/10.5555/3495724.3495883.
- Choudhary, Tavishi. 2024. AI Bias Audit Tool. Cyber Smart Teens. Accessible at: https://cybersmartteens.wixsite.com/aibiasaudit.
- Chris Smith, Brian McGuire, Ting Huang, and Gary Yang. 2006. The History of Artificial Intelligence. Course paper, CSEP 590A: History of Computing, University of Washington. Available at: https://courses.cs.washington.edu/courses/csep590/06au/projects/history-ai.pdf.
- De Vynck, Gerrit. 2023. ChatGPT leans liberal, research shows. The Washington Post. Available at: https://www.washingtonpost.com/technology/2023/08/16/chatgpt-ai-political-bias-research.
- Durach, C.F.; Kembro, J.; Wieland, A. A New Paradigm for Systematic Literature Reviews in Supply Chain Management. J. Supply Chain Manag. 2017, 53, 67–85. [Google Scholar] [CrossRef]
- Eubanks, Virginia. 2018. Automating Inequality: How High-Tech Tools Profile, Police, and Punish the Poor. St. Martin's Press: Picador.
- Getahun, Hannah. 2023. CHATGPT Could Be Used for Good, but like Many Other AI Models, It's Rife with Racist and Discriminatory Bias. Insider. Available at: https://www.insider.com/chatgpt-is-like-many-other-ai-models-rife-with-bias-2023-1.
- Gillis, Alexander S., and Mary K. Pratt. 2023. What Is Machine Learning Bias?: Definition from Whatis. Enterprise AI. TechTarget. Available at: https://www.techtarget.com/searchenterpriseai/definition/machine-learning-bias-algorithm-bias-or-AI-bias.
- Gross, N. What ChatGPT Tells Us about Gender: A Cautionary Tale about Performativity and Gender Biases in AI. Soc. Sci. 2023, 12, 435. [Google Scholar] [CrossRef]
- Huyue Zhang, Angela, et al. 2022. The Four Domains of Global Platform Governance. Centre for International Governance Innovation. Available at: https://www.cigionline.org/publications/the-four-domains-of-global-platform-governance.
- Jacob Livingston Slosser. 2021. Artificial Intelligence. In: The Routledge Handbook of Law and Society, edited by Mariana Valverde et al.
- Jindal, A. Misguided Artificial Intelligence: How Racial Bias is Built Into Clinical Models. Brown J. Hosp. Med. 2022, 2. [Google Scholar] [CrossRef]
- Jurafsky, D., and J. H. Martin. 2019. Speech and Language Processing: An Introduction to Natural Language Processing, Computational Linguistics, and Speech Recognition. 3rd ed. Available at: https://web.stanford.edu/~jurafsky/slp3/.
- Kapczynski, A. Data and Democracy: An Introduction. The Knight Institute, Columbia University. 2021. Available online: https://knightcolumbia.org/content/data-and-democracy-an-introduction (accessed on 21 October 2021).
- Khieu, K., and N. Narwal. 2018. CS224N: Detecting and Classifying Toxic Comments. Available at: https://web.stanford.edu/class/archive/cs/cs224n/cs224n.1184/reports/6837517.pdf.
- Kitchenham, B. 2004. Procedures for performing systematic reviews. Keele, UK: Keele University, School of Computer Science and Mathematics, Software Engineering Group.
- Krasadakis, George. 2023. The Ethical Concerns Associated with the General Adoption of AI. Medium, 60 Leaders. Available at: https://medium.com/60-leaders/the-ethical-concerns-associated-with-the-general-adoption-of-ai-ab893e9b5196.
- Lazaro, Gina. 2022. Understanding Gender and Racial Bias in AI. ALI Social Impact Review, Advanced Leadership Initiative. Available at: https://www.sir.advancedleadership.harvard.edu/articles/understanding-gender-and-racial-bias-in-ai.
- Wikipedia. . . List of Ethnic Slurs. Wikimedia Foundation. 2024. Available online: https://en.wikipedia.org/wiki/List_of_ethnic_slurs (accessed on 4 April 2024).
- Lorè, Filippo, et al. 2023. An AI Framework to Support Decisions on GDPR Compliance. Shibboleth Authentication Request. [CrossRef]
- McLeod, Juanita. 2021. Understanding Racial Terms and Differences. Available at: https://www.edi.nih.gov/blog/communities/understanding-racial-terms-and-differences.
- Manyika, James, Jake Silberg, and Brittany Presten. 2019. What Do We Do About the Biases in AI?. Harvard Business Review. Available at: https://hbr.org/2019/10/what-do-we-do-about-the-biases-in-ai.
- Noble, S. U. 2018. Algorithms of Oppression: How Search Engines Reinforce Racism. New York: New York University Press.
- O'Neil, C. 2016. Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy. New York: Crown.
- Pichai, Sundar. 2023. An important next step on our AI journey. CEO of Google and Alphabet. Available at: https://blog.google/technology/ai/bard-google-ai-search-updates.
- Radford, A. 2019. Language Models are Unsupervised Multitask Learners. Technical report, OpenAI. Available at: https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf.
- Radford, A. 2018. Improving Language Understanding by Generative Pre-training. Technical report, OpenAI. Available at: https://s3-us-west-2.amazonaws.com/openai-assets/research-covers/language-unsupervised/language_understanding_paper.pdf.
- Ruiz, N.G. Discrimination Experiences Shape Most Asian Americans' Lives. Pew Research Center Race & Ethnicity. 2023. Available online: https://www.pewresearch.org/race-ethnicity/2023/11/30/discrimination-experiences-shape-most-asian-americans-lives/ (accessed on 30 November 2023).
- Serrano, Jody. 2024. ChatGPT AI Racism Study on African-American English. Gizmodo. Available at: https://qz.com/chatgpt-ai-racism-study-african-american-english-1851324423.
- Sexton, Nick Kim. 2015. Study Reveals Americans' Subconscious Racial Biases. NBCNews.Com, NBCUniversal News Group. Available at: https://www.nbcnews.com/news/asian-america/new-study-exposes-racial-preferences-americans-n413371.
- Shanmugavadivel, K. , et al. 2022. Deep learning based sentiment analysis and offensive language identification on multilingual code-mixed data. Scientific Reports 12, Article number: 21900. [CrossRef]
- Sutaria, Niral, CISA, ACA. 2022. Bias and Ethical Concerns in Machine Learning. ISACA Journal. Accessible at: https://www.isaca.org/resources/isaca-journal/issues/2022/volume-4/bias-and-ethical-concerns-in-machine-learning.
- Tang, Yu-Chien, et al. 2023. Customer Intent Detection via Agent Response Contrastive and Generative Pre-Training. arXiv:2310.09773. [CrossRef]
- U.S. Equal Employment Opportunity Commission. (n.d.). Significant EEOC race/color cases (covering private and federal sectors). Home | U.S. Equal Employment Opportunity Commission. Accessible at: https://www.eeoc.
- Wagner, P. 2021. Data Privacy - The Ethical, Sociological, and Philosophical Effects of Cambridge Analytica. Available at SSRN: https://ssrn. 3782. [Google Scholar]
- Weed, Mike. 2006. Sports Tourism Research 2000–2004: A Systematic Review of Knowledge and a Meta-Evaluation of Methods. Journal of Sport & Tourism 11(1): 5-30. [CrossRef]
- West, Darrell M. 2023. Comparing Google Bard with OpenAI's ChatGPT on political bias, facts, and morality. Brookings Institution. Accessible at: https://www.brookings.edu/articles/comparing-google-bard-with-openais-chatgpt-on-political-bias-facts-and-morality.
- Wu, X.; et al. 2022. A survey of human-in-the-loop for machine learning. Future Generation Computer Systems 135: 364-381. [CrossRef]
- Zack, Travis et al. 2023. Coding Inequity: Assessing GPT-4's Potential for Perpetuating Racial and Gender Biases in Healthcare. medRxiv, Cold Spring Harbor Laboratory Press. Available at: https://www.medrxiv.org/content/10.1101/2023.07.13.23292577v2.
- Zhu, S.; Yu, T.; Xu, T.; Chen, H.; Dustdar, S.; Gigan, S.; Gunduz, D.; Hossain, E.; Jin, Y.; Lin, F.; et al. Intelligent Computing: The Latest Advances, Challenges, and Future. Intell. Comput. 2023, 2, 006. [Google Scholar] [CrossRef]











Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).