Submitted:
25 April 2025
Posted:
25 April 2025
You are already at the latest version
Abstract
Keywords:
1. Introduction
1.1. Background
1.2. Research Questions
2. Methodology
3. United States
3.1. ChatGPT (OpenAI)
3.2. Claude (Anthropic)
3.3. Gemini (Google DeepMind)
3.4. LLaMA (Meta)
3.5. Azure and Copilot Integration (Microsoft)
3.6. Grok (xAI)
3.7. Titan and Nova Models (Amazon AWS)
3.8. Nemotron Ultra (Nvidia and Core Partners)
3.9. Perplexity AI: Retrieval-Augmented Generation (RAG) at Scale
4. China
4.1. DeepSeek (DeepSeek AI)
4.2. Kimi (Moonshot AI)
4.3. ERNIE Bot (Baidu)
4.4. Tongyi Qianwen (Alibaba)
4.5. Doubao & Cloud Lark (ByteDance)
4.6. PanGu Series (Huawei)
4.7. Hunyuan (Tencent)
4.8. Spark Model (iFlytek)
4.9. Baichuan Model Family (Baichuan Intelligence)
4.10. ChatGLM Series (Zhipu AI & Tsinghua University)
4.11. Zi Dong Tai Chu (Chinese Academy of Sciences)
4.12. InternLM (SenseTime)
4.13. ABAB & MiniMax Models (MiniMax AI)
5. European Union
5.1. BLOOM (France): The BigScience Open Multilingual Model
5.2. Aleph Alpha’s Luminous (Germany): Efficient Multilingual AI for Europe
5.3. Mistral AI (France): Small, High-Performance Models with Open Release
5.4. OpenGPT-X (Germany/EU Consortium): Large Models for European Sovereignty
6. United Kingdom
6.1. DeepMind’s Gopher and Chinchilla (UK): Pioneering Large-Scale and Efficient LLMs
6.2. Stability AI’s Open-Source LLMs (UK): Democratizing Access
7. India
7.1. BharatGPT (CoRover)
7.2. Airavata (Hindi LLM by AI4Bharat)
7.3. IndicGPT (AI4Bharat IIT Madras)
7.4. IndicBERT (AI4Bharat Multilingual ALBERT)
8. Japan
8.1. Nekomata Series (Rinna)
8.2. AI Scientist Project (Sakana AI)
8.3. Governmental Research (NICT)
9. South Korea
9.1. HyperCLOVA (NAVER)
9.2. Multimodal Open-Source Innovation (Kakao)
9.3. EXAONE (LG AI Research)
10. Canada
10.1. Command and Aya Series (Cohere)
10.2. Academic and Research Contributions: Mila, Vector Institute, and More
11. Other Notable Countries
11.1. AI21 Labs (Israel)
11.2. GigaChat (Russia)
11.3. Falcon (UAE & Saudi Arabia)
11.4. Research and Policy Initiatives by CSIRO Data61 (Australia)
11.5. Bode (Brazil and Latin America)
11.6. InkubaLM (Africa)
12. Discussion
12.1. U.S. LLMs: Technical Leadership Anchored by Scale and Infrastructure
12.2. Compute Dominance: GPU Access as a Strategic Enabler
12.3. China: Innovation Under Hardware Constraints
12.4. Open Source and Regional AI Sovereignty: Europe and Beyond
12.5. Open Source vs. Commercial Use
12.6. Multilingualism and Societal Relevance as Strategic Differentiators
12.7. National Capacity and the Structural Foundations of LLM Advancement
12.8. Algorithmic Advances as an Entry Point for Small Teams
13. Conclusions
References
- A. J. Thirunavukarasu, D. S. J. Ting, K. Elangovan, L. Gutierrez, T. F. Tan, and D. S. W. Ting, ‘Large language models in medicine’, Nature medicine, vol. 29, no. 8, pp. 1930–1940, 2023. [CrossRef]
- W. C. Choi, I. C. Choi, C. I. Chang, and L. C. Lam, ‘Comparison of Claude (Sonnet and Opus) and ChatGPT (GPT-4, GPT-4o, GPT-o1) in Analyzing Educational Image-based Questions from Block-Based Programming Assessments’, in 2025 14th International Conference on Information and Education Technology (ICIET), IEEE, 2025.
- W. C. Choi and C. I. Chang, ‘Enhancing Education with ChatGPT 4o and Microsoft Copilot: A Review of Opportunities, Challenges, and Student Perspectives on LLM-Based Text-to-Image Generation Models’, 2025.
- I. Chang, W. C. Choi, and I. C. Choi, ‘Challenges and Limitations of Using Artificial Intelligence Generated Content (AIGC) with ChatGPT in Programming Curriculum: A Systematic Literature Review’, in Proceedings of the 2024 7th Artificial Intelligence and Cloud Computing Conference, 2024.
- I. Chang, W. C. Choi, and I. C. Choi, ‘A Systematic Literature Review of the Opportunities and Advantages for AIGC (OpenAI ChatGPT, Copilot, Codex) in Programming Course’, in Proceedings of the 2024 7th International Conference on Big Data and Education, 2024.
- W. C. Choi, J. Peng, I. C. Choi, H. Lei, L. C. Lam, and C. I. Chang, ‘Improving Young Learners with Copilot: The Influence of Large Language Models (LLMs) on Cognitive Load and Self-Efficacy in K-12 Programming Education’, in Proceedings of the 2025 International Conference on Artificial Intelligence and Education (ICAIE), Suzhou, China, 2025.
- R. Aghaei et al., ‘Harnessing the Potential of Large Language Models in Modern Marketing Management: Applications, Future Directions, and Strategic Recommendations’, arXiv preprint arXiv:2501.10685, 2025. [CrossRef]
- P. Homoki and Z. Z\Hodi, ‘Large language models and their possible uses in law’, Hungarian Journal of Legal Studies, vol. 64, no. 3, pp. 435–455, 2024. [CrossRef]
- Culver, P. Hicks, M. Milenkovic, S. Shanmugavelu, and T. Becker, ‘Scientific computing with large language models’, arXiv preprint arXiv:2406.07259, 2024. [CrossRef]
- Reuters, ‘China’s Tencent debuts large language AI model, says open for enterprise use’. [Online]. Available: https://www.reuters.com/technology/chinas-tencent-says-large-language-ai-model-hunyuan-available-enterprise-use-2023-09-07/.
- BigScience, ‘BigScience Large Open-science Open-access Multilingual Language Model’. 2025. [Online]. Available: https://huggingface.co/bigscience/bloom.
- OpenAI, ‘OpenAI’. 2025. [Online]. Available: https://openai.com/.
- L. Ouyang et al., ‘Training language models to follow instructions with human feedback’, Advances in neural information processing systems, vol. 35, pp. 27730–27744, 2022.
- J. Achiam et al., ‘Gpt-4 technical report’, arXiv preprint arXiv:2303.08774, 2023. [CrossRef]
- Anthropic, ‘Anthropic’. [Online]. Available: https://www.anthropic.com/.
- Anthropic, ‘Claude’. 2025. [Online]. Available: https://claude.ai.
- Google, ‘Google DeepMind’. 2025. [Online]. Available: https://deepmind.google/.
- A. Chowdhery et al., ‘Palm: Scaling language modeling with pathways’, Journal of Machine Learning Research, vol. 24, no. 240, pp. 1–113, 2023.
- R. Anil et al., ‘Palm 2 technical report’, arXiv preprint arXiv:2305.10403, 2023. [CrossRef]
- G. Team et al., ‘Gemini: a family of highly capable multimodal models’, arXiv preprint arXiv:2312.11805, 2023. [CrossRef]
- Meta, ‘Meta AI’. 2025. [Online]. Available: https://www.meta.ai/.
- H. Touvron et al., ‘Llama: Open and efficient foundation language models’, arXiv preprint arXiv:2302.13971, 2023. [CrossRef]
- Microsoft, ‘Microsoft AI’. 2025. [Online]. Available: https://microsoft.ai/.
- Microsoft, ‘Microsoft Azure: Cloud Computing Services’. 2025. [Online]. Available: https://azure.microsoft.com/.
- J. Stratton, ‘An introduction to microsoft copilot’, in Copilot for Microsoft 365: Harness the Power of Generative AI in the Microsoft Apps You Use Every Day, Springer, 2024, pp. 19–35.
- xAI, ‘xAI’. 2025. [Online]. Available: https://x.ai/.
- xAI, ‘Grok 3 Beta — The Age of Reasoning Agents’. [Online]. Available: https://x.ai/news/grok-3.
- Amazon Web Services, ‘Artificial Intelligence (AI) on AWS - AI Technology’. 2025. [Online]. Available: https://aws.amazon.com/ai/.
- Amazon Web Services, ‘Amazon Titan Foundation Models’. [Online]. Available: https://aws.amazon.com/tw/bedrock/amazon-models/titan/.
- Amazon Web Services, ‘Amazon Nova Foundation Models’. [Online]. Available: https://aws.amazon.com/ai/generative-ai/nova/.
- NVIDIA, ‘NVIDIA Llama Nemotron’. [Online]. Available: https://www.nvidia.com/en-us/ai-data-science/foundation-models/llama-nemotron/.
- Vespa, ‘We Make AI Work at Perplexity – Delivering AI-Powered Search at Scale’. [Online]. Available: https://vespa.ai/perplexity/.
- Rest of World, ‘DeepSeek and chip bans have supercharged AI innovation in China’. [Online]. Available: https://restofworld.org/2025/china-ai-boom-chip-ban-deepseek/.
- DeepSeek, ‘DeepSeek’. 2025. [Online]. Available: https://www.deepseek.com/.
- W. C. Choi and C. I. Chang, ‘Advantages and Limitations of Open-Source Versus Commercial Large Language Models (LLMs): A Comparative Study of DeepSeek and OpenAI’s ChatGPT’, 2025.
- Wikipedia, ‘DeepSeek’. 2025. [Online]. Available: https://en.wikipedia.org/wiki/DeepSeek.
- Nature News, ‘China’s cheap, open AI model DeepSeek thrills scientists’, Nature, 2025. [CrossRef]
- X. Bi et al., ‘Deepseek llm: Scaling open-source language models with longtermism’, arXiv preprint arXiv:2401.02954, 2024. [CrossRef]
- DeepSeek, ‘deepseek-ai (DeepSeek)’. 2025. [Online]. Available: https://huggingface.co/deepseek-ai.
- Reuters, ‘China’s DeepSeek releases AI model upgrade, intensifies rivalry with OpenAI’. [Online]. Available: https://www.reuters.com/technology/artificial-intelligence/chinas-deepseek-releases-ai-model-upgrade-intensifies-rivalry-with-openai-2025-03-25/.
- Moonshot AI, ‘Kimi - Moonshot AI’. 2025. [Online]. Available: https://kimi.moonshot.cn/.
- Global Times, ‘China’s Moonshot AI fuels domestic large model app frenzy, aiming to overtake ChatGPT’. [Online]. Available: https://www.globaltimes.cn/page/202403/1309421.shtml.
- South China Morning Post, ‘Alibaba-backed Moonshot AI claims breakthrough in expanded Chinese-character prompt for Kimi chatbot’. [Online]. Available: https://www.scmp.com/tech/big-tech/article/3256109/alibaba-backed-moonshot-ai-claims-breakthrough-expanded-chinese-character-prompt-kimi-chatbot.
- South China Morning Post, ‘Chinese AI start-up Moonshot cuts LLM feature price amid fierce domestic competition’. [Online]. Available: https://www.scmp.com/tech/tech-trends/article/3273619/chinese-ai-start-moonshot-cuts-llm-feature-price-amid-fierce-domestic-competition.
- K. Team et al., ‘Kimi k1. 5: Scaling reinforcement learning with llms’, arXiv preprint arXiv:2501.12599, 2025. [CrossRef]
- South China Morning Post, ‘Chinese AI start-up MiniMax releases low-cost open-source models that rival top chatbots’. [Online]. Available: https://www.scmp.com/tech/big-tech/article/3294900/chinese-ai-start-minimax-releases-low-cost-open-source-models-rival-top-chatbots.
- Wikipedia, ‘ERNIE Bot’. 2025. [Online]. Available: https://en.wikipedia.org/wiki/Ernie_Bot.
- Y. Sun et al., ‘Ernie: Enhanced representation through knowledge integration’, arXiv preprint arXiv:1904.09223, 2019. [CrossRef]
- A. Velinov, ‘Chinese LLMs vs Western LLMs - Developments, Comparisons, and Global Outlook’. [Online]. Available: https://www.linkedin.com/pulse/chinese-llms-vs-western-developments-comparisons-global-velinov-sqeyf/.
- Reuters, ‘Alibaba upgrades AI model Tongyi Qianwen, releases industry-specific models’. [Online]. Available: https://www.reuters.com/technology/alibaba-upgrades-ai-model-tongyi-qianwen-releases-industry-specific-models-2023-10-31/.
- Alibaba Cloud, ‘Tongyi Qianwen - Alibaba Cloud’. 2025. [Online]. Available: https://tongyi.aliyun.com/.
- UC Today, ‘What is Tongyi Qianwen? Alibaba’s ChatGPT Rival’. [Online]. Available: https://www.uctoday.com/collaboration/what-is-tongyi-qianwen-alibabas-chatgpt-rival/.
- CNBC, ‘Alibaba opens AI model Tongyi Qianwen to public in competition with Baidu, SenseTime’. [Online]. Available: https://www.cnbc.com/2023/08/30/alibaba-opens-ai-model-tongyi-qianwen-to-public.html.
- R. Zheng et al., ‘Secrets of rlhf in large language models part i: Ppo’, arXiv preprint arXiv:2307.04964, 2023. [CrossRef]
- ByteDance, ‘ByteDance’. 2025. [Online]. Available: https://www.bytedance.com/.
- ByteDance, ‘Doubao AI Tool’. 2025. [Online]. Available: https://www.doubao.com/chat/.
- China Internet Watch, ‘ByteDance’s AI Surge: Introducing the Doubao Model Family’. [Online]. Available: https://www.ciw.news/p/bytedance-ai-surge-doubao.
- Wikipedia, ‘Doubao chatbot’. 2025. [Online]. Available: https://zh.wikipedia.org/wiki/%E8%BE%93%E5%8C%85_(%E8%81%8A%E5%A4%A9%E6%9C%BA%E5%99%A8%E4%BA%BA).
- South China Morning Post, ‘TikTok owner ByteDance launches low-cost Doubao AI models for enterprises, initiating a price war in crowded mainland market’. [Online]. Available: https://sc.mp/2h1kd?utm_source=copy-link&utm_campaign=3262781&utm_medium=share_widget.
- TMTPost, ‘Alibaba and ByteDance Intensify AI Race Despite Rule Tightening’. [Online]. Available: https://en.tmtpost.com/post/6485054.
- X. Ren et al., ‘Pangu-Sigma: Towards trillion parameter language model with sparse heterogeneous computing’, arXiv preprint arXiv:2303.10845, 2023. [CrossRef]
- Wikipedia, ‘Huawei PanGu’. 2025. [Online]. Available: https://en.wikipedia.org/wiki/Huawei_PanGu.
- Counterpoint Research, ‘Huawei Expands its AI Ambition With Pangu Large Models’. [Online]. Available: https://www.counterpointresearch.com/insights/huawei-expands-its-ai-ambition-with-pangu-large-models/.
- Huawei, ‘AI for Good at HAS 2024: GenAI for Modern Drug Discovery’. [Online]. Available: https://www.huawei.com/en/huaweitech/industry-trends/has24-takeaways-pangu-omdia.
- 1AI, ‘The Chinese Academy of Sciences’ self-developed AI model “Zidong Taichu 3.0” was released in the first half of this year to optimize intelligent driving training’. [Online]. Available: https://www.1ai.net/en/4953.html.
- Tencent, ‘Tencent Hunyuan’. 2025. [Online]. Available: https://hunyuan.tencent.com/.
- X. Sun et al., ‘Hunyuan-large: An open-source moe model with 52 billion activated parameters by tencent’, arXiv preprint arXiv:2411.02265, 2024. [CrossRef]
- Analytics India Magazine, ‘Tencent Launches Hunyuan Large, Outperforms Llama 3.1 70B & 405B’. [Online]. Available: https://analyticsindiamag.com/ai-news-updates/tencent-launches-hunyuan-large-outperforms-llama-3-1-70b-405b/.
- iFLYTEK, ‘iFLYTEK Spark Model’. 2025. [Online]. Available: https://xinghuo.xfyun.cn/.
- Baidu Baike, ‘iFLYTEK Spark Cognitive Model’. 2025. [Online]. Available: https://baike.baidu.com/item/%E8%AE%AF%E9%A3%9E%E6%98%9F%E7%81%AB%E8%AE%A4%E7%9F%A5%E5%A4%A7%E6%A8%A1%E5%9E%8B/62912838.
- Live Science, ‘New Chinese AI model “better than industry leader” in key metrics’. [Online]. Available: https://www.livescience.com/technology/artificial-intelligence/chinese-ai-model-spark-35-better-than-open-ai-gpt4.
- China Daily, ‘iFlytek unveils upgraded LLM that “outperforms GPT-4 Turbo”’. [Online]. Available: https://www.chinadaily.com.cn/a/202406/28/WS667e67f6a31095c51c50b617.html.
- Wikipedia, ‘Baichuan’. 2025. [Online]. Available: https://en.wikipedia.org/wiki/Baichuan.
- AI Fun, ‘Baichuan Intelligence’. [Online]. Available: https://www.aifun.cc/en/sites/baichuanzhineng.html.
- A. Yang et al., ‘Baichuan 2: Open large-scale language models’, arXiv preprint arXiv:2309.10305, 2023. [CrossRef]
- AI Fun, ‘Bai Xiaoying’. [Online]. Available: https://www.aifun.cc/en/sites/baixiaoying.html.
- Zhipu AI, ‘ChatGLM’. 2025. [Online]. Available: https://chatglm.cn/.
- SiliconANGLE, ‘Chinese AI startup Zhipu raises $300M in funding’. [Online]. Available: https://siliconangle.com/2023/10/20/chinese-ai-startup-zhipu-raises-300m-funding/.
- Nature News, ‘China’s ChatGPT: why China is building its own AI chatbots’, Nature, 2024. [CrossRef]
- T. GLM et al., ‘Chatglm: A family of large language models from glm-130b to glm-4 all tools’, arXiv preprint arXiv:2406.12793, 2024. [CrossRef]
- C. Liu et al., ‘CPMI-ChatGLM: parameter-efficient fine-tuning ChatGLM with Chinese patent medicine instructions’, Scientific reports, vol. 14, no. 1, p. 6403, 2024. [CrossRef]
- H. Zhou et al., ‘GLMLog: Log Anomaly Detection Method Based on ChatGLM’, in 2024 10th International Conference on Computer and Communications (ICCC), IEEE, 2024, pp. 1923–1927.
- Institute of Automation, Chinese Academy of Sciences, ‘Zi Dong Tai Chu’. 2025. [Online]. Available: https://taichu-web.ia.ac.cn/.
- Baidu Baike, ‘Zi Dong Tai Chu’. 2025. [Online]. Available: https://baike.baidu.com/item/%E7%B4%AB%E4%B8%9C%E5%A4%AA%E5%88%9D/57969205.
- Chinese Academy of Sciences, ‘Chinese Academy of Sciences Launches Its Next-generation AI Model’. [Online]. Available: https://english.cas.cn/newsroom/cas_media/202306/t20230620_332105.shtml.
- Team, ‘Internlm: A multilingual language model with progressively enhanced capabilities’. 2023.
- The Decoder, ‘With “InternLM”, China enters the race for large language models’. [Online]. Available: https://the-decoder.com/with-internlm-china-enters-the-race-for-large-language-models/.
- China Daily, ‘SenseTime’s large language model service open for registration’. [Online]. Available: https://global.chinadaily.com.cn/a/202308/31/WS64eff802a31035260b81f363.html.
- InternLM Team, ‘InternLM/InternLM: Official release of InternLM series’. 2025. [Online]. Available: https://github.com/InternLM/InternLM.
- Yicai Global, ‘SenseTime, Shanghai AI Lab Debut New-Generation LLM’. [Online]. Available: https://www.yicaiglobal.com/news/chinas-sensetime-three-other-institutes-debut-new-gen-llm.
- MiniMax, ‘MiniMax’. 2025. [Online]. Available: https://www.minimaxi.com/.
- Wikipedia, ‘MiniMax’. 2025. [Online]. Available: https://en.wikipedia.org/wiki/MiniMax_(company).
- Wikipedia contributors, ‘BLOOM (language model)’. 2025. [Online]. Available: https://en.wikipedia.org/wiki/BLOOM_(language_model).
- BigScience, ‘Introducing The World’s Largest Open Multilingual Language Model - BLOOM’. 2025. [Online]. Available: https://bigscience.huggingface.co/.
- T. Le Scao et al., ‘Bloom: A 176b-parameter open-access multilingual language model’, 2023.
- Aleph Alpha, ‘Aleph Alpha’. 2025. [Online]. Available: https://aleph-alpha.com/.
- Aleph Alpha, ‘Luminous’. 2025. [Online]. Available: https://www.luminous.com/.
- Wikipedia, ‘Aleph Alpha’. 2025. [Online]. Available: https://en.wikipedia.org/wiki/Aleph_Alpha.
- adesso, ‘Quickstart with a European-based large language model: Aleph Alpha’s “Luminous”’. [Online]. Available: https://www.adesso.de/en/news/blog/quickstart-with-a-european-based-large-language-model-aleph-alphas-luminous.jsp.
- Aleph Alpha, ‘Aleph Alpha Luminous Benchmarks’. 2025. [Online]. Available: https://aleph-alpha.com/wp-content/uploads/Performance-Report_Luminous_Aleph-Alpha.pdf.
- Aleph Alpha, ‘Luminous: European AI closes gap to world leaders’. [Online]. Available: https://aleph-alpha.com/luminous-european-ai-closes-gap-to-world-leaders/.
- Techzine Europe, ‘Aleph Alpha is releasing two open-source models fully compliant with the European AI Act’. [Online]. Available: https://www.techzine.eu/blogs/privacy-compliance/123863/aleph-alphas-open-source-llms-fully-comply-with-the-ai-act/.
- Mistral AI, ‘Mistral AI’. 2025. [Online]. Available: https://mistral.ai/.
- A. Q. Jiang et al., ‘Mistral 7B’, 2023. [Online]. Available: https://arxiv.org/abs/2310.06825. [CrossRef]
- Mistral AI, ‘Announcing Mistral 7B’. [Online]. Available: https://mistral.ai/news/announcing-mistral-7b.
- Mistral AI, ‘Mistral NeMo’. [Online]. Available: https://mistral.ai/news/mistral-nemo.
- Mistral AI, ‘Models Overview’. 2025. [Online]. Available: https://docs.mistral.ai/getting-started/models/models_overview/.
- OpenGPT-X Consortium, ‘OpenGPT-X, Multilingual. Open. European’. 2025. [Online]. Available: https://opengpt-x.de/.
- A. Herten and S. Kesselheim, ‘Multilingual and Open Source: OpenGPT-X Releases Large Language Model’. [Online]. Available: https://www.fz-juelich.de/en/news/archive/announcements/2024/multilingual-and-open-source-opengpt-x-releases-large-language-model.
- The Register, ‘Germany to host Europe’s first exascale supercomputer’. [Online]. Available: https://www.theregister.com/2022/06/16/first_european_exascale/.
- LUMI Supercomputer, ‘Open LLMs for transparent AI in Europe’. [Online]. Available: https://www.lumi-supercomputer.eu/open-euro-llm/.
- DeepMind, ‘Language modelling at scale: Gopher, ethical considerations, and retrieval’. [Online]. Available: https://deepmind.google/discover/blog/language-modelling-at-scale-gopher-ethical-considerations-and-retrieval/.
- W. Rae et al., ‘Scaling language models: Methods, analysis & insights from training gopher’, arXiv preprint arXiv:2112.11446, 2021. [CrossRef]
- DeepMind, ‘An empirical analysis of compute-optimal large language model training’. [Online]. Available: https://deepmind.google/discover/blog/an-empirical-analysis-of-compute-optimal-large-language-model-training/.
- Hoffmann et al., ‘Training compute-optimal large language models’, arXiv preprint arXiv:2203.15556, 2022. [CrossRef]
- Stability AI, ‘Stability AI’. 2025. [Online]. Available: https://stability.ai/.
- Wikipedia, ‘Stability AI’. 2025. [Online]. Available: https://en.wikipedia.org/wiki/Stability_AI.
- A. Alford, ‘Stability AI Open-Sources 7B Parameter Language Model StableLM’. [Online]. Available: https://www.infoq.com/news/2023/05/stablelm-release/.
- CoRover Pvt. Ltd, ‘BharatGPT’. [Online]. Available: https://corover.ai/bharatgpt/.
- Analytics India Magazine, ‘CoRover.ai’s BharatGPT Surpasses 2,000 Downloads on Hugging Face Within Days’. [Online]. Available: https://analyticsindiamag.com/ai-news-updates/corover-ais-bharatgpt-surpasses-2000-downloads-on-hugging-face-within-days/.
- CoRover, ‘CoRover/BharatGPT-3B-Indic’. [Online]. Available: https://huggingface.co/CoRover/BharatGPT-3B-Indic.
- Computer Weekly, ‘Indian large language models gain momentum’. [Online]. Available: https://www.computerweekly.com/news/366575652/Indian-large-language-models-gain-momentum.
- AI4Bharat, ‘ai4bharat/Airavata’. [Online]. Available: https://huggingface.co/ai4bharat/Airavata.
- J. Gala et al., ‘Airavata: Introducing hindi instruction-tuned llm’, arXiv preprint arXiv:2401.15006, 2024. [CrossRef]
- The Economic Times, ‘AI4Bharat releases Hindi LLM “Airavata”’. [Online]. Available: https://economictimes.indiatimes.com/tech/technology/ai4bharat-releases-hindi-llm-airavata/articleshow/107146024.cms.
- IndicGPT, ‘aashay96/indic-gpt’. 2025. [Online]. Available: https://huggingface.co/aashay96/indic-gpt.
- AI4Bharat, ‘AI4Bharat’. [Online]. Available: https://ai4bharat.iitm.ac.in/.
- Brijeshkumar Y. Panchal, ‘Examine the Opportunities and Challenges of Large Language Model (LLM) For Indic Languages’, jisem, vol. 10, no. 26s, pp. 301–325, Mar. 2025. [CrossRef]
- AI4Bharat, ‘ai4bharat/indic-bert’. [Online]. Available: https://huggingface.co/ai4bharat/indic-bert.
- Kakwani et al., ‘IndicNLPSuite: Monolingual corpora, evaluation benchmarks and pre-trained multilingual language models for Indian languages’, in Findings of the association for computational linguistics: EMNLP 2020, 2020, pp. 4948–4961.
- Rinna Co., Ltd, ‘Rinna Co., Ltd’. 2025. [Online]. Available: https://rinna.co.jp/global/.
- Rinna Co., Ltd, ‘Rinna Releases “Nekomata” Series: Japanese Continued Pre-trained Models Based on Qwen’. [Online]. Available: https://rinna.co.jp/news/2023/12/20231221.html.
- Rinna Co., Ltd., ‘Japanese-Language AI Models Based on Tongyi Qianwen (Qwen) Were Launched by rinna’. Feb. 2023. [Online]. Available: https://www.alibabacloud.com/blog/rinna-launched-ai-models-trained-in-the-japanese-language—nekomata-series—based-on-alibaba-clouds-qwen-models_600719.
- Rinna Co., Ltd., ‘rinna/bilingual-gpt-neox-4b-minigpt4’. 2023. [Online]. Available: https://huggingface.co/rinna/bilingual-gpt-neox-4b-minigpt4.
- Sakana AI, ‘The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery’. [Online]. Available: https://sakana.ai/ai-scientist/.
- Sakana AI, ‘Sakana AI Wins Award at US-Japan Competition for Defense Innovation’. Mar. 24, 2025. [Online]. Available: https://sakana.ai/defense-challenge-2025/.
- T. Akiba, M. Shing, Y. Tang, Q. Sun, and D. Ha, ‘Evolutionary optimization of model merging recipes’, Nature Machine Intelligence, pp. 1–10, 2025. [CrossRef]
- National Institute of Information and Communications Technology, ‘National Institute of Information and Communications Technology’. 2025. [Online]. Available: https://www.nict.go.jp/en/.
- KDDI Corporation, ‘KDDI’. 2025. [Online]. Available: https://www.kddi.com/english/.
- Telecompaper, ‘KDDI and NICT Start Joint Research on Multimodal LLMs’. Jul. 2024. [Online]. Available: https://www.telecompaper.com/news/kddi-and-nict-start-joint-research-on-multimodal-llms–1505017.
- L.-J. Consortium, ‘LLM-jp’. 2023. [Online]. Available: https://llm-jp.nii.ac.jp/en/resources/.
- A. Aizawa et al., ‘Llm-jp: A cross-organizational project for the research and development of fully open japanese llms’, arXiv preprint arXiv:2407.03963, 2024. [CrossRef]
- NAVER Corporation, ‘HyperCLOVA X’. 2025. [Online]. Available: https://clova.ai/en/hyperclova.
- B. Kim et al., ‘What changes can large-scale language models bring? intensive study on hyperclova: Billions-scale korean generative pretrained transformers’, arXiv preprint arXiv:2109.04650, 2021. [CrossRef]
- M. Yoo et al., ‘Hyperclova x technical report’, arXiv preprint arXiv:2404.01954, 2024. [CrossRef]
- NAVER Corporation, ‘CLOVA Studio’. 2025. [Online]. Available: https://clova.ai/en/clova-studio.
- NAVER Corporation, ‘CLOVA X’. 2025. [Online]. Available: https://clova-x.naver.com/welcome.
- Kakao Corporation, ‘Kakao’. 2025. [Online]. Available: https://www.kakaocorp.com/.
- J. Cha, W. Kang, J. Mun, and B. Roh, ‘Honeybee: Locality-enhanced projector for multimodal llm’, in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 13817–13827.
- LG AI Research, ‘LG AI Research’. 2025. [Online]. Available: https://www.lgresearch.ai/.
- LG AI Research, ‘EXAONE Official Site’. 2025. [Online]. Available: https://www.lgresearch.ai/exaone.
- Research et al., ‘EXAONE 3.5: Series of Large Language Models for Real-world Use Cases’, arXiv preprint arXiv:2412.04862, 2024. [CrossRef]
- Cohere, ‘Cohere’s Command R+ Model’. 2025. [Online]. Available: https://docs.cohere.com/v2/docs/command-r-plus.
- Cohere, ‘Introducing Aya’. 2025. [Online]. Available: https://cohere.com/research/aya.
- A. Üstün et al., ‘Aya model: An instruction finetuned open-access multilingual language model’, arXiv preprint arXiv:2402.07827, 2024. [CrossRef]
- Cohere, ‘Cohere Aya 101’. [Online]. Available: https://huggingface.co/CohereLabs/aya-101.
- Cohere, ‘Aya Expanse - 8B’. 2025. [Online]. Available: https://huggingface.co/CohereLabs/aya-expanse-8b.
- Cohere, ‘Aya Expanse - 32B’. 2025. [Online]. Available: https://huggingface.co/CohereLabs/aya-expanse-32b.
- Mila, ‘Mila – Quebec AI Institute’. [Online]. Available: https://mila.quebec/.
- AI21 Labs, ‘AI21 Labs’. [Online]. Available: https://www.ai21.com/.
- AI21 Labs, ‘Jurassic-1 Jumbo’. [Online]. Available: https://llmmodels.org/tools/jurassic-1-jumbo/.
- Lieber, O. Sharir, B. Lenz, and Y. Shoham, ‘Jurassic-1: Technical details and evaluation’, White Paper. AI21 Labs, vol. 1, no. 9, pp. 1–17, 2021.
- AI21 Editorial Team, ‘Announcing Jurassic-2 and Task-Specific APIs’. [Online]. Available: https://www.ai21.com/blog/introducing-j2/.
- Sberbank, ‘GigaChat’. [Online]. Available: https://giga.chat/.
- Yandex, ‘YaLM-100B: Pretrained language model with 100B parameters’. [Online]. Available: https://github.com/yandex/YaLM-100B.
- Yandex, ‘Yandex publishes YaLM 100B, the largest GPT-like neural network in open source’. [Online]. Available: https://yandex.com/company/press_center/press_releases/2022/2022-23-06.
- Technology Innovation Institute, ‘tiiuae (Technology Innovation Institute)’. [Online]. Available: https://huggingface.co/tiiuae.
- Inception AI, ‘JAIS’. [Online]. Available: https://inceptionai.ai/jais/index.html.
- C. Data61, ‘Artificial Intelligence Foundation Models’, CSIRO, Feb. 2024. [Online]. Available: https://www.csiro.au/-/media/D61/Files/2400012DATA61REPORTAIFoundationModelsWEB240208-1.pdf.
- L. Garcia et al., ‘Introducing bode: a fine-tuned large language model for Portuguese prompt-based task’, arXiv preprint arXiv:2401.02909, 2024. [CrossRef]
- L. AI, ‘Lelapa AI launches Africa’s first AI large language model’. [Online]. Available: https://africaworld.princeton.edu/news/2024/lelapa-ai-launches-africa%E2%80%99s-first-ai-large-language-model.
| Company | LLM Series | Representative Models | Open-Source or Commercial Use |
Characteristics |
|---|---|---|---|---|
| OpenAI | ChatGPT | GPT-3, GPT-3.5, GPT-4, GPT-o1, GPT-o3 | Commercial Use | One of the earliest and leading LLM developers; GPT-3 introduced scaling laws; GPT-3.5 added RLHF; GPT-4 is multimodal; from GPT-o1 onward, strong image recognition and generation capabilities. |
| Anthropic | Claude | Claude 1, Claude 2, Claude 3 | Commercial Use | Focused on safety and alignment via Constitutional AI; partial transparency; delivered via API. |
| Google DeepMind | Gemini | Palm, Gemini 1, Gemini 1.5 | Commercial Use | Built on Pathways architecture; optimized for reasoning and multimodal tasks; deployed via Bard and Vertex AI. |
| Meta | LLaMA | LLaMA 1, LLaMA 2 |
Open-Source | LLaMA 2 released under a permissive license; emphasizes transparency; used in research and industry. |
| Microsoft | Copilot | GPT-based (via OpenAI) |
Commercial Use | Integrates OpenAI models into Office and enterprise tools; also acts as cloud and platform provider. |
| xAI | Grok | Grok 1, Grok 1.5 |
Commercial Use | Focuses on real-time information retrieval; integrated with X (formerly Twitter); emphasizes responsiveness and open discourse. |
| Amazon AWS | Titan / Nova | Titan Text G1 – Express, Titan Embeddings, Nova Pro, Nova Sonic | Commercial Use | Delivered through Bedrock; designed for enterprise needs; supports text generation, embeddings, image/video generation, and speech-based tasks; emphasizes modularity, privacy, and flexibility. |
| Nvidia | Nemotron Ultra | Nemotron-4 340B, Nemotron-3 43B | Open-Access (with constraints) | Reference models for GPU optimization; supports alignment and fine-tuning; delivered via NeMo framework. |
| Perplexity AI | Perplexity | Perplexity 2, Perplexity 3 | Commercial Use | Retrieval-augmented generation (RAG); supports live citation and source tracking; optimized for factual consistency. |
| Company | LLM Series | Representative Models | Open-Source or Commercial Use |
Characteristics |
|---|---|---|---|---|
| DeepSeek AI | DeepSeek | DeepSeek-V3, DeepSeek-R1 | Open-Source | High performance with low training cost; bilingual (Chinese and English); Mixture-of-Experts architecture; top-ranked in Chinese benchmarks |
| Moonshot AI | Kimi | Kimi 1.0, Kimi k1.5 | Commercial Use (Partially Open) | Ultra-long context (up to 2M characters); designed for enterprise use; supports multimodal inputs |
| Baidu | ERNIE | ERNIE 3.5, ERNIE 4.0 |
Commercial Use | Knowledge-enhanced pretraining; strong Chinese focus; integrated in search, cloud, and mobile apps |
| Alibaba | Tongyi Qianwen | Tongyi 1.0, 2.0; Qwen-7B/14B | Mixed (Open + Commercial) | Bilingual models; large-scale proprietary + open-source Qwen; widely deployed in Alibaba’s ecosystem; frequently used in performance comparisons |
| ByteDance | Doubao | Doubao-Pro, Doubao-Lite | Commercial Use | Long context (128k tokens); aggressive pricing; integrated in Douyin and Feishu (Lark); strong in image generation |
| Huawei | PanGu | PanGu-Alpha, PanGu-Σ, PanGu 5.0 | Commercial Use | Trillion-scale parameters; MoE structure; tailored for verticals (weather, finance, manufacturing); cloud integration |
| Tencent | Hunyuan | Hunyuan-VL | Commercial Use (Partially Open) | Dense Transformer; strong performance on Chinese tasks; integrated in WeChat, QQ, and fintech tools |
| iFlytek | SparkDesk | Spark v2.0–v4.0+ | Commercial Use | NLP + speech capabilities; strong in education; supports text/audio |
| Baichuan Intelligence | Baichuan | Baichuan-7B, 13B, 53B, Baichuan 3/4 | Open-Source | Apache license; bilingual; strong community adoption; high performance in Chinese; widely used in research |
| Zhipu AI | ChatGLM | ChatGLM-6B, ChatGLM2, ChatGLM-130B | Open-Source + API | Open-source and API accessible; bilingual; strong Chinese NLP; widely used in industry and academia |
| CAS | Zi Dong Tai Chu (ZDTC) | ZDTC 1.0, 2.0 | Research Access Only | Multimodal model; supports text, vision, audio, and video; deployed in robotics and legal AI |
| SenseTime & Shanghai AI Lab | InternLM | InternLM-123B, InternLM2.0, InternLM-7B/20B | Mixed (Open + Commercial) | Long-context (300k chars); strong reasoning; multilingual; used in SenseChat and enterprise APIs |
| MiniMax AI | ABAB / MiniMax | ABAB-6.5, MiniMax-Text-01, MiniMax-VL-01 | Mixed (Open + Commercial) | MoE-based; multimodal and multilingual; competitive open models and aggressive deployment |
| Company | LLM Series | Representative Models | Open-Source or Commercial Use |
Characteristics |
|---|---|---|---|---|
| France | BLOOM | BLOOM 176B | Open-Source (Responsible AI License) | Multilingual (46 languages); trained on public compute; early open 100B+ model; strong research transparency |
| Germany | Luminous | Luminous-base, Luminous-supreme (70B) | Commercial API (partial open) | European-language focus; explainability tools; GDPR-compliant cloud deployment; strong efficiency |
| France | Mistral | Mistral 7B, Mistral NeMo 12B |
Open-Source | Strong performance at small scale; Grouped-Query & Sliding Window Attention; efficient long context handling |
| Germany / EU Consortium | OpenGPT-X | Teuken-7B | Open-Source | Trained on EU supercomputers; 24 EU languages; public infrastructure use; compliant with EU AI Act goals |
| Company | LLM Series | Representative Models | Open-Source or Commercial Use |
Characteristics |
|---|---|---|---|---|
| DeepMind | Gopher, Chinchilla | Gopher (280B), Chinchilla (70B) | Research Only | Gopher explored scaling and risks; Chinchilla introduced compute-optimal training; strong impact on efficiency paradigms |
| Stability AI | StableLM | StableLM Alpha 3B, 7B | Open-Source | Trained on 1.5T tokens; code + text generation; open weights; aligns with open-science and low-barrier deployment principles |
| Company | LLM Series | Representative Models | Open-Source or Commercial Use |
Characteristics |
|---|---|---|---|---|
| CoRover.ai | BharatGPT | BharatGPT 3B | Open-Source | Multilingual; 12+ Indian languages; optimized for edge/offline deployment; deployed in real-world apps like AskDisha |
| AI4Bharat (IIT Madras) | Airavata | Airavata v0.1 | Open-Source | Hindi instruction-following model; fine-tuned from OpenHathi (LLaMA-2 based); licensed for reuse |
| AI4Bharat | IndicGPT | IndicGPT (GPT-2 based) | Open-Source | GPT-2 decoder trained on Indic texts; supports Hindi, Tamil, Bengali, Telugu; strong open-ended generation |
| AI4Bharat | IndicBERT | IndicBERT, IndicBERT v2 | Open-Source | ALBERT-based encoder model; multilingual (12 Indic languages); high efficiency and performance on NLU tasks |
| Company | LLM Series | Representative Models | Open-Source or Commercial Use |
Characteristics |
|---|---|---|---|---|
| Rinna | Nekomata | Nekomata 7B, Nekomata 14B | Open-Source (based on Qwen) | Japanese-focused models fine-tuned from Qwen; support for complex grammar; optimized for local NLP tasks |
| Sakana AI | EvoLLM-JP, EvoVLM-JP | EvoLLM-JP, AI Scientist | Research & Evaluation Stage | Nature-inspired model merging; multimodal and bilingual capabilities; focus on automation and low-resource training |
| NICT + KDDI | LLM-Japan (LLM-jp) | LLM-jp-13B | In Development / Research | Multimodal models for text + images; national initiative; large-scale training on Japanese web data; hallucination mitigation focus |
| Company | LLM Series | Representative Models | Open-Source or Commercial Use |
Characteristics |
|---|---|---|---|---|
| NAVER | HyperCLOVA | HyperCLOVA (204B), HyperCLOVA X | Commercial + API Access | GPT-3 scale; trained on Korean-heavy data; integrated into search, assistants, enterprise plugins |
| Kakao Brain | Honeybee | Honeybee Multimodal Module | Open-Source (Module) | Adds vision-language capabilities to LLMs; interprets images; extensible for other models |
| LG AI Research | ExaONE | ExaONE-32B, ExaONE-7.8B | Mixed (Public Access for Smaller Models) | Multimodal reasoning; text + vision; Korean-optimized; smaller variants open for research |
| Company | LLM Series | Representative Models | Open-Source or Commercial Use |
Characteristics |
|---|---|---|---|---|
| Cohere | Command, Aya | Command R+, Aya 101, Aya Expanse 8B/32B | Mixed (Commercial + Open) | Instruction-tuned for enterprise; Aya supports 100+ languages; focus on underrepresented languages and ethical training |
| Mila / Vector Institute | N/A (Research Contributions) | Contributed to BLOOM, multilingual benchmarks | Research Only | Focus on low-resource adaptation, sustainability, and LLM ethics; key roles in multilingual and open-access initiatives |
| Country / Region |
LLM Series / Project | Representative Models | Open-Source or Commercial Use |
Characteristics |
|---|---|---|---|---|
| Israel | Jurassic | Jurassic-1 Jumbo, Jurassic-2 | Commercial API | Early GPT-3-scale model; strong in English and Hebrew; among first non-U.S. LLMs of this scale |
| Russia | GigaChat, YaLM | GigaChat, YaLM-100B | Mixed (GigaChat API, YaLM Open) | YaLM-100B was the largest openly licensed model at its time; Russian-centric language focus; national compute use |
| UAE | Falcon, Jais | Falcon 40B, 180B; Jais 13B/30B | Open-Source | Falcon 40B/180B trained on 1T+ tokens; Jais optimized for Arabic-English tasks; strategic positioning in global AI |
| Saudi Arabia | Noor | Noor 10B, AraGPT-2 | Research Only | Arabic-centric; KAUST and SDAIA involved; early contributions to regional LLM capacity |
| Australia | (Research & Policy) | No major native LLMs | Policy/Research Contributions | Focus on responsible AI, policy, and inclusion of Indigenous languages; active university involvement |
| Brazil / LatAm | Bode | Bode 7B, 13B (LLaMA-based) | Open-Source | Portuguese-focused; tailored LLaMA models; increasing local NLP tool development |
| Africa | InkubaLM | InkubaLM, SafaBERT | Open-Source (Small-scale) | Supports multiple African languages; foundation for future scaling; driven by Lelapa AI and Masakhane community |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).