Preprint
Review

This version is not peer-reviewed.

Tracing the Evolution of Artificial Intelligence: A Review of Tools, Frameworks, and Technologies (1950–2025)

Submitted:

09 November 2025

Posted:

11 November 2025

You are already at the latest version

Abstract
Artificial Intelligence (AI) has evolved remarkably over the past seven decades, transforming from simple rule-based systems into complex multimodal and generative frameworks capable of reasoning, creativity, and perception. This review traces the chronological development of AI tools, highlighting key milestones that shaped the field—from the early symbolic programs like Logic Theorist and ELIZA to the emergence of modern large-scale models such as GPT-4, Gemini, and Claude. The study explores the progression across distinct eras: the foundational period of symbolic reasoning (1940s–1970s), the rise of machine learning and statistical modeling (1980s–2000s), the deep learning revolution (2010s), and the recent explosion of generative and multimodal systems (2020–2025). Each phase reflects a major shift in how intelligence is defined, represented, and implemented—from handcrafted logic to data-driven learning and now to context-aware multimodal understanding. By reviewing over fifty significant AI tools and frameworks, this paper provides a comprehensive overview of how incremental innovations in computation, data availability, and model architecture have collectively enabled the current state of AI. The work concludes with insights on how this evolution paves the way for the next generation of agentic and real-time AI systems capable of seamless interaction across text, image, audio, and video modalities.
Keywords: 
;  ;  ;  ;  

1. Introduction

Artificial Intelligence (AI) has become one of the most transformative forces of the 21st century, shaping industries, research, and everyday life in ways that were once thought impossible. From simple logic-based programs created in university labs to today’s generative systems capable of producing human-like text, images, and even videos, the journey of AI represents not just technological progress but also humanity’s evolving understanding of intelligence itself. Each era of AI—from symbolic reasoning and rule-based systems to neural networks, deep learning, and now multimodal generative models—has introduced new ways of learning, interacting, and problem-solving.
The purpose of this paper is to provide a comprehensive and chronological review of the evolution of AI tools and frameworks, capturing how each innovation built upon the previous generation to form the foundation of modern artificial intelligence. While many studies focus on specific branches of AI such as computer vision, natural language processing, or robotics, few have presented a complete timeline connecting the earliest symbolic systems to the current multimodal and agentic era. This review aims to fill that gap by examining over seven decades of AI development—from early expert systems like MYCIN and DENDRAL to advanced generative models like GPT-4, Gemini, and Sora.
Writing this paper in 2025 holds particular significance. The world is now witnessing a rapid convergence of text, image, speech, and video understanding through multimodal AI models, making it essential to reflect on how we arrived at this point and what lessons past innovations can teach us. The pace of AI progress often makes older technologies seem obsolete, yet many of today’s breakthroughs are deeply rooted in the principles and experiments of earlier decades. Understanding this historical evolution helps researchers, educators, and practitioners appreciate the foundations of current AI tools and foresee the directions future technologies might take.
Ultimately, this paper serves as both a historical reflection and an educational guide, offering readers a unified perspective on how AI tools have matured from isolated programs into interconnected, intelligent ecosystems. By mapping this evolution, the study underscores the continuity of innovation that defines artificial intelligence—not as a sudden revolution, but as a long, collaborative journey of human curiosity and technological advancement.

2. Early Foundations (1940s–1970s)

2.1. Logic Theorist (1956)

The Logic Theorist, developed in 1956 by Allen Newell, Herbert A. Simon, and Cliff Shaw at the RAND Corporation, is widely recognized as the first artificial intelligence (AI) program purposely designed to mimic human problem-solving and reasoning skills. Its primary goal was to prove mathematical theorems, specifically those found in the seminal work Principia Mathematica by Alfred North Whitehead and Bertrand Russell. The program demonstrated that a machine could engage in automated reasoning by successfully proving 38 of the first 52 theorems in chapter two of Principia Mathematica. Remarkably, the Logic Theorist not only replicated human logic but also discovered new, sometimes shorter, proofs for some theorems, exemplifying the ability of AI to surpass human intellectual processes in specific domains [1].
The development of the Logic Theorist was grounded in symbolic logic and heuristic search—a method that applies rules of thumb to explore possible solutions in a vast search space. The program represented logical expressions symbolically and searched through combinations of these expressions to construct proofs. It used heuristics to efficiently navigate the tree of possible inferences, thus avoiding an exhaustive and computationally prohibitive search through all branches. This heuristic approach was inspired by how humans tackle complex problems, prioritizing promising paths based on certain criteria rather than brute-force search [2].
Architecturally, the Logic Theorist consisted primarily of two components: a knowledge base containing axioms and previously proven theorems, and an inference engine that applied logical rules to generate new theorems. The proof search was conducted by generating sub-proofs for propositions through symbolic manipulations. Although programming languages and computational resources were very primitive at the time, the team implemented the Logic Theorist using handwritten cards that were distributed among collaborators and later run on computers available at RAND [3]. This development also led to the creation of the Information Processing Language (IPL), which facilitated list processing and symbolic computation—precursors to programming languages such as LISP, which became fundamental in AI research [4].
Mathematically, the Logic Theorist operated within propositional calculus as formulated in Principia Mathematica. Its proofs often utilized methods such as proof by contradiction, where the algorithm would assume the negation of a theorem and derive contradictions to verify its truth. By starting from axioms and applying inference rules, the program mechanized the deductive reasoning process typical of mathematical proof. The discovery of novel proofs for certain theorems, such as theorem 2.85, highlighted its capability not only to replicate but to improve upon human mathematical reasoning [5].
The impact of the Logic Theorist on AI and cognitive science was profound. It marked the first practical demonstration that machines could perform high-order intellectual tasks involving symbolic reasoning. It laid foundational concepts such as heuristic search, symbolic processing, and the notion that reasoning could be mechanized, influencing subsequent AI programs and research. Herbert Simon and Allen Newell, the principal developers, were later recognized with Turing Awards for their pioneering contributions to AI and computer science. The Logic Theorist remains a landmark in AI history, illustrating the power of symbolic AI and heuristic-driven problem solving in the early quest to imitate human thought through machines

2.2. ELIZA (1966)

ELIZA, developed in 1966 by Joseph Weizenbaum at MIT, is one of the earliest and most influential natural language processing programs and is widely regarded as the first chatbot to simulate human conversation [6]. Unlike later conversational agents built on complex machine learning models, ELIZA operated fundamentally through pattern matching and substitution rules, using scripts designed to give the illusion of understanding without true comprehension. The most famous script, known as "DOCTOR," simulated a Rogerian psychotherapist by reflecting users’ statements back to them in the form of questions, encouraging users to continue dialogue . For instance, when a user expressed feelings of sadness or worry, ELIZA would respond with inquiries prompting further reflection, employing simple yet effective mechanisms to seem empathetic and human-like.
ELIZA’s architecture was based on symbolic processing where user inputs were analyzed for keywords or phrases, and then matching scripted patterns were triggered to generate predefined responses. It did not involve semantic or contextual understanding, which distinguished it from genuine human communication. Nevertheless, the program’s simplistic style led to what became known as the "ELIZA effect," [7] where users attributed human-like understanding and emotions to the program despite its mechanical nature. This phenomenon highlighted important psychological and philosophical questions about human interaction with machines and the nature of intelligence. Interestingly, Weizenbaum himself was surprised and troubled by how readily some users, including his own secretary, formed emotional connections with ELIZA, underscoring the powerful illusion of understanding it created.
The development of ELIZA marked a significant milestone in artificial intelligence and human-computer interaction because it demonstrated that computers could engage users in seemingly meaningful conversation through language-based interfaces. Its creation also helped lay the groundwork for research in natural language understanding, dialogue systems, and chatbot development. ELIZA’s influence extended beyond technological innovation; it sparked critical debates about machine intelligence, ethics, and the limitations of AI, since the program could simulate empathy without actual cognitive or emotional processing. Weizenbaum stressed that ELIZA was a tool, not a sentient entity, cautioning against overestimating what machines can replicate in terms of human thought and feeling [8].
In summary, ELIZA represents a pioneering experiment in AI programming that combined early symbolic processing techniques and heuristic scripting to simulate conversation. While mathematically it did not engage in logical deduction, it excelled in procedural language pattern matching, which was enough to make a lasting impact on AI, natural language processing, and cognitive science. Its legacy continues in modern chatbots that have since evolved with more sophisticated models, but ELIZA remains a foundational artifact demonstrating the potentials and boundaries of early conversational AI [9]

2.3. SHRDLU (1970)

SHRDLU, developed between 1968 and 1970 by Terry Winograd at MIT as part of his PhD thesis, represents a pioneering natural language understanding system that could interpret and execute commands within a simplified virtual environment known as the "blocks world." This virtual world consisted of various geometric shapes—blocks, pyramids, and boxes of different colors—that SHRDLU manipulated in response to user instructions phrased in natural English. The program could perform tasks such as moving objects, stacking blocks, and answering questions about the state of the environment, such as “What is on the table?” or “Is there a pyramid on the block?” The user interacted with SHRDLU through a dialogue interface that allowed both commands and queries, making it an early example of an interactive AI system capable of understanding and reasoning about language and its referents in a controlled world [10].
From a technical standpoint, SHRDLU combined syntax parsing, semantic interpretation, and a form of procedural knowledge representation to understand instructions and generate responses. It was implemented mainly in Lisp and Micro Planner on a DEC PDP-6 computer. The system parsed English commands, mapped them onto internal logical representations of objects and their properties in the blocks world [11] , and then used rules to manipulate these objects or generate language responses. This architecture showcased an integrated approach to natural language processing, combining syntax, semantics, and pragmatics within a finite domain. SHRDLU’s knowledge base was dynamic, maintaining an up-to-date model of the world’s state, which it could modify as it executed actions or learned new commands. The logical foundation often related to first-order logic, where objects, properties, and relations were explicitly represented, enabling reasoning over the set of facts.
Beyond its technical achievements, SHRDLU critically demonstrated the challenges and potential of language understanding in AI. Winograd’s work exposed the brittleness of symbolic systems when faced with the complexity and ambiguity of natural language outside a restricted domain. Yet, SHRDLU’s ability to follow complex instructions and engage in clarifying dialogues to resolve ambiguities was groundbreaking, highlighting the possibility of interactive AI systems that understand human language contextually. It influenced subsequent research in computational linguistics, human-computer interaction, and AI, laying a foundation for modern dialogue systems, virtual assistants, and robotic control by natural language. SHRDLU’s legacy extends into contemporary natural language processing by inspiring approaches to syntactic and semantic integration and knowledge representation in AI, demonstrating how procedural representations can bridge language and action within a defined world framework[12].
This comprehensive understanding of SHRDLU’s development, architecture, functionality, and impact should serve well for your research paper, situating it as a landmark system in natural language understanding and AI history.

2.4. DENDRAL (1965–1970)

Developed between 1965 and 1970 at Stanford University by Edward Feigenbaum, Bruce Buchanan, Joshua Lederberg, and Carl Djerassi, DENDRAL is recognized as the first successful expert system and a landmark in artificial intelligence history. Designed primarily as a chemical analysis tool, DENDRAL aimed to assist organic chemists in identifying unknown molecular structures by analyzing mass spectrometry data. Unlike prior AI systems, DENDRAL encoded expert knowledge in chemistry as heuristic rules and used this domain-specific expertise to automate the process of hypothesis generation and evaluation—effectively emulating the decision-making process of skilled human chemists. This enabled the system not only to generate possible molecular structures but also to predict their corresponding mass spectra and compare them with the experimental data to select the most plausible hypotheses [13].
The architecture of DENDRAL was characterized by a clear separation between its knowledge base and inference engine, forming a paradigm later foundational to expert systems. Its knowledge base contained encoded chemical heuristics and domain rules, while the inference engine orchestrated the generation and evaluation of candidate molecular structures through a heuristic search process. The system consisted mainly of two programs: Heuristic Dendral, which performed structure elucidation using domain-specific heuristics, and Meta-Dendral, which learned new rules by analyzing patterns in chemical data, pioneering early forms of machine learning. Technically, DENDRAL combined symbolic reasoning with heuristic search strategies to prune the vast search space of chemical structures, applying constraints and domain knowledge to improve computational efficiency and accuracy [14].
Mathematically, DENDRAL relied on combinatorial and graph-theoretic methods to represent molecular structures as cyclic graphs and trees, aligning with its etymological roots (“dendron” meaning tree in Greek). It employed algorithms that enumerated possible configurations constrained by chemical valence rules and mass spectral data, effectively narrowing down the candidates through heuristic pruning. By translating chemical knowledge into formal rules and representing candidate molecules structurally, DENDRAL bridged symbolic AI with practical chemical problem-solving.
The significance of DENDRAL extends beyond its application, as it pioneered the concept and successful implementation of expert systems—computer programs that encapsulate expert-level knowledge and reasoning ability within a specific domain. It demonstrated AI’s potential to enhance scientific discovery and decision-making by explicitly codifying expert heuristics into programmable knowledge bases. Furthermore, DENDRAL influenced subsequent AI developments, including MYCIN and other domain-specific expert systems, and helped establish research paradigms in AI knowledge engineering and human-computer interaction. Its success bolstered funding and interest in AI during the 1970s and fundamentally shaped the trajectory of knowledge-based systems in computer science and expert decision support [15].
This detailed overview captures DENDRAL’s impact as a pioneering expert system in chemical analysis through heuristic symbolic reasoning, foundational to AI’s evolution in expert knowledge representation and automated problem-solving.

2.5. MYCIN (1972)

MYCIN, developed in the early 1970s at Stanford University by Edward Shortliffe and colleagues, is a pioneering rule-based expert system designed to assist physicians in diagnosing bacterial infections and recommending appropriate antibiotic treatments. Originating from the Stanford Heuristic Programming Project, MYCIN employed a knowledge base of approximately 600 production rules encoding the clinical decision-making expertise of infectious disease specialists. The system operated through a backward chaining inference engine that interactively queried physicians with a series of simple yes/no or text-based questions about the patient’s symptoms and lab results. Based on the gathered data, MYCIN produced a ranked list of potential causative bacteria with associated confidence levels and provided tailored antibiotic therapy suggestions, adjusting dosages based on patient-specific factors such as body weight [16].
Technically, MYCIN’s architecture distinguished itself by clearly separating its inference procedures from its domain-specific knowledge, embodying a modular approach that became foundational for later expert systems. The system implemented a certainty factor model to handle uncertainty in medical diagnosis, representing the confidence in rules and conclusions, although this model was heuristic rather than strictly Bayesian due to computational and practical constraints. MYCIN was also notable for its explanatory capabilities, where it could justify its recommendations by tracing back through the rules and questions that led to its conclusions, addressing concerns about transparency and trust in AI systems for medical decision-making [17].
From a mathematical and logical perspective, MYCIN relied on production rules—if-then statements representing clinical knowledge—and heuristic-driven backward chaining to efficiently search the space of diagnostic possibilities. The certainty factor calculus combined evidential strengths from multiple rules to arrive at probabilistic-like confidence measures in diagnoses and recommendations. While these certainty factors lacked rigorous statistical foundations, they provided a practical framework for handling medical diagnostic uncertainty.
MYCIN’s impact on AI and medicine was profound, demonstrating that expert knowledge in specialized domains could be formalized and leveraged by computer systems to perform tasks at the level of human specialists. Its success inspired numerous subsequent expert systems and contributed significantly to the development of knowledge engineering, rule-based reasoning, and explanation facilities in AI. MYCIN also sparked discussions on the ethical and practical implications of AI in clinical settings, particularly regarding the role of decision support versus autonomous diagnosis. Although constrained by the computational power and data availability of its time, MYCIN laid the groundwork for modern clinical decision support systems and remains a seminal example of early AI applied to real-world expertise [18].
This comprehensive account captures MYCIN’s development, architecture, operation, mathematical underpinnings, and its lasting significance as a foundation for rule-based AI systems, making it a critical milestone in the history of artificial intelligence and medical informatics.

2.6. PROLOG (1972)

Prolog, created in 1972 by Alain Colmerauer and Philippe Roussel at Aix-Marseille University in France, is a logic programming language foundational to artificial intelligence and natural language processing tasks. The name Prolog derives from the French phrase "Programmation en Logique," meaning programming in logic. Rooted in first-order predicate logic, Prolog allows programmers to express knowledge declaratively, using facts and rules rather than explicit procedural code. The language’s execution is driven by a goal-directed search process implementing resolution theorem proving with Horn clauses, which it processes through a mechanism called backward chaining. This enables Prolog to perform symbolic reasoning, automated theorem proving, and natural language understanding, making it particularly suited for AI applications [19].
The language emerged from combining two efforts: Alain Colmerauer’s focus on natural language processing and Robert Kowalski’s theoretical work on the procedural interpretation of logic. The original Prolog system included an interpreter written in Algol-W and later versions influenced by David H.D. Warren, who developed the Warren Abstract Machine (WAM), a standard virtual machine architecture for efficient Prolog implementation. This contributed significantly to Prolog’s widespread adoption and the establishment of the Edinburgh syntax standard that most Prolog implementations follow. Prolog’s logical foundation centers on Horn clauses that define relations, with computations being queries that the interpreter attempts to satisfy by proving them from the known facts and rules [20].
Prolog’s computational model is distinguished by features like unification, backtracking, and negation as failure, which together allow it to explore the space of potential solutions effectively. Mathematically, Prolog programs can be viewed as logic formulas in clausal form and the resolution proof method underpins its operation. The declarative style facilitates expressing complex information and constraints naturally, enabling applications in expert systems, theorem proving, language parsing, knowledge representation, and automated planning.
Historically, Prolog gained international prominence during the 1980s particularly with the Japanese Fifth Generation Computer Systems project, which sought to leverage logic programming for advanced AI on parallel computers. While Prolog faced competition and technical challenges, its conceptual clarity and powerful expressiveness secured its lasting place in AI programming. Today, Prolog remains a prominent symbolic programming language, well-suited to tasks demanding sophisticated pattern matching, reasoning, and rule-based logic processing, continuing to underpin research and applications in artificial intelligence and computational linguistics [21].

2.7. LISP (1958)

LISP, created in 1958 by John McCarthy at MIT, is one of the oldest and most influential programming languages in the field of artificial intelligence. McCarthy designed LISP with the explicit goal of providing a powerful and flexible language specifically suited for AI research, particularly symbolic manipulation and processing of list structures. The language introduced several groundbreaking concepts, such as recursive functions, dynamic typing, and automatic memory management through garbage collection. Its primary data structure, the list, and its associated operations—car and cdr—allowed natural and efficient expression of symbolic computation, which is central to AI [22].
Initially, McCarthy proposed a notation called "M-expressions" to make LISP more readable, but it was quickly abandoned in favor of "S-expressions," a more straightforward parenthesized prefix notation that became the hallmark of LISP’s syntax. The first LISP interpreter was implemented by Steve Russell on an IBM 704, surprising McCarthy by demonstrating that the eval function—a core evaluation procedure—could be directly compiled into machine code, enabling practical execution. LISP’s design was mathematically rooted in the theory of recursive functions and lambda calculus, which provided a formal foundation for defining computable functions and reasoning about process execution [23].
Over the decades, LISP became the dominant AI programming language because of its unparalleled ability to handle symbolic information, its extensibility through macros, and its interactive development environment. It powered pioneering AI projects such as SHRDLU and expert systems, and influenced other languages like Scheme and Common Lisp. LISP introduced concepts critical to AI, including symbolic reasoning, code as data (homoiconicity), and advanced macro systems, which helped AI researchers experiment with new ideas efficiently. Common Lisp, standardized in the 1980s and 1990s, unified various dialects and remains widely used in AI, research, and industry.
LISP’s mathematical essence lies in its recursive function definitions, symbolic lists, and evaluation semantics, which collectively enabled it to express complex AI algorithms elegantly. Its legacy continues to permeate AI research, functional programming, and symbolic computation, making it a foundational technology in the history and development of artificial intelligence [24].
This overview encapsulates LISP’s historical context, innovative design principles, mathematical underpinnings, and its critical role as the dominant AI programming language for decades.

3. The Rise of Machine Learning (1980s–2000s)

3.1. SOAR (1983)

SOAR, developed in 1983 by John Laird, Allen Newell, and Paul Rosenbloom at Carnegie Mellon University, is a pioneering cognitive architecture designed to model human-like reasoning and general intelligence. Rooted deeply in cognitive science and artificial intelligence, SOAR aims to create fixed computational building blocks for general intelligent agents capable of performing a wide array of tasks such as decision-making, problem-solving, planning, and natural language understanding. The architecture embodies a unified theory of cognition, evolving from Allen Newell’s Problem Space Hypothesis—a foundational AI theory stating that all goal-directed behavior can be framed as search within a space of possible states. SOAR operationalizes this by continuously selecting and applying operators to change an agent’s state, analogous to how humans approach problems step-by-step [25].
SOAR’s architecture integrates procedural memory (knowledge of how to do things) with working memory (representation of the current situation), enabling a dynamic cognitive cycle. Procedural knowledge is encoded as if-then production rules (condition-action pairs) that match against the contents of working memory. Unlike other systems, SOAR fires all matching rules in parallel, allowing concurrent context-dependent retrieval of knowledge. When the system encounters an impasse—lacking knowledge to proceed—SOAR automatically creates a substate that recursively applies the same problem-solving process, leading to the generation of subgoals and hierarchical task decomposition. This universal subgoaling approach naturally models complex cognitive behaviors including planning and learning. The working memory itself is structured as a symbolic graph rooted in the current state, facilitating flexible representation of knowledge [26].
Mathematically, SOAR’s computation is grounded in state-transition systems, production rule matching, and search algorithms navigating complex problem spaces. The production system supports parallel rule matching while ensuring that behavior is decomposed into primitive operator applications approximating human reaction times ( 5 ~ 0ms per step). The substate recursion forms a sophisticated mathematical structure supporting hierarchical problem-solving, setting SOAR apart from prior architectures. Its design reflects the interplay of symbolic AI with psychological realism, striving to unify AI capabilities and cognitive modeling [27].
Over the decades, SOAR has evolved into a comprehensive cognitive architecture widely employed in AI research to build intelligent agents and model human behavior in a range of applications—from robotics to gaming to natural language processing. Maintained today by John Laird’s research group at the University of Michigan, SOAR remains a seminal reference point for cognitive architectures, continuing to influence theories of general intelligence by bridging formal computational models with empirical cognitive science insights.
This detailed exploration covers SOAR’s theoretical foundations, developmental history, architecture, mathematical underpinnings, and its significance in AI and cognitive science [28].

3.2. ALVINN (1989)

ALVINN (Autonomous Land Vehicle In a Neural Network), developed in 1989 at Carnegie Mellon University by Dean Pomerleau, is a landmark early neural network system designed for autonomous vehicle navigation and road following. As one of the first practical applications of artificial neural networks to self-driving cars, ALVINN was built around a three-layer backpropagation network that took as input a combination of sensory data from a forward-facing camera and a laser range finder, processing images and depth information to produce steering commands for the vehicle. The architecture comprised roughly 1217 input units, a hidden layer with 29 units, and an output layer with 46 units dedicated to representing a range of possible steering angles, along with feedback units to integrate temporal context [29].
Trained initially on simulated road images, ALVINN used supervised learning with backpropagation to optimize its weights, enabling it to predict the required steering angle to follow the road accurately. The training leveraged data augmentation techniques, generating shifted versions of images to robustly handle diverse driving scenarios and prevent the system from failing when deviating from the road centerline. Operationally, ALVINN was deployed on the NAVLAB autonomous test vehicle, demonstrating its ability to navigate complex outdoor environments—including varied weather conditions—at speeds competitive with traditional computer vision approaches of the time. This success illustrated the promise of adaptive, data-driven models in autonomous navigation, surpassing rigid rule-based systems by learning representations tailored dynamically to sensory inputs.
Mathematically [30], ALVINN employed a feedforward neural network with supervised gradient descent optimization (backpropagation) to minimize prediction errors between the network output and actual steering commands. The network model embodied non-linear function approximation, capturing the complex mapping from high-dimensional sensory input spaces (images and range data) to control actions. The use of feedback units introduced an element of temporal memory, allowing the network to incorporate information about the recent past into its decision-making process—related conceptually to what would later be developed as recurrent neural architectures.
ALVINN’s introduction was a key milestone in AI and robotics, illustrating how neural networks could be harnessed for real-world control problems under uncertainty and noise. It heralded the feasibility of end-to-end learning for vehicle control, influencing future developments in autonomous driving technology and machine learning-driven robotics. The system highlighted the importance of sensor fusion, adaptive learning, and robust training strategies in autonomous navigation, and its design principles form the basis for many modern AI-based self-driving systems [31].

3.3. Backpropagation Algorithm (1986)

The backpropagation algorithm, popularized in 1986 by David Rumelhart, Geoffrey Hinton, and Ronald Williams, revolutionized the training of artificial neural networks by providing an efficient method to compute gradients of the loss function with respect to weights and biases for multilayered networks. Prior to this breakthrough, training deep networks was computationally infeasible due to the difficulty in attributing errors to internal layers. Backpropagation operationalizes gradient descent in a highly scalable way, allowing error signals from the output layer to be propagated backward through the network layers by systematically applying the chain rule of calculus [32].
The algorithm works in two main phases during each training iteration. In the forward pass, the network computes the output by propagating inputs through successive layers via weighted connections and activation functions. In the backward pass, the algorithm calculates the gradient of the loss function relative to each weight by propagating the error backward from the output layer to the input layer. These gradients are then used to update network parameters in the direction that minimizes the output error. Mathematically, for a weight w j k l connecting neuron k in layer l 1 to neuron j in layer l, the update step corresponds to computing the partial derivative of the cost C with respect to that weight, C w j k l , and using it in gradient descent:
w j k l w j k l η C w j k l
where η is the learning rate. This process is repeated iteratively over examples, applying the chain rule to efficiently allocate error contributions among all weights.
Backpropagation’s significance lies not only in its algorithmic efficiency but also in its enabling of modern deep learning. It permits neural networks to learn internal representations automatically, making them capable of solving highly complex and nonlinear tasks including image recognition, speech processing, and natural language understanding. The 1986 paper by Rumelhart, Hinton, and Williams marked a turning point in AI because it provided a practical and scalable learning algorithm that forms the foundation of nearly all contemporary deep learning architectures. Since then, backpropagation has become the workhorse of neural network training, supported by abundant optimizations, regularization methods, and variants that have further expanded the reach of AI solutions [33].
In essence, backpropagation combines deep mathematical concepts from calculus with computational strategies to effectively optimize multilayer neural networks, ushering in the era of powerful, adaptive artificial intelligence systems.

3.4. WEKA (1993)

WEKA (Waikato Environment for Knowledge Analysis), developed starting in 1993 at the University of Waikato in New Zealand, is a comprehensive, open-source machine learning and data mining software suite widely used for teaching, research, and practical data analysis. Created by a team led by Professor Ian H. Witten and colleagues, WEKA was born from a desire to provide an accessible platform that integrates a large collection of machine learning algorithms with tools for data preprocessing, visualization, and evaluation, all accessible through user-friendly graphical interfaces. Originally, WEKA was developed with agricultural and horticultural data applications in mind, but it has since evolved to support a diverse array of domains and data types [34].
WEKA’s architecture centers around modular components that perform key data mining tasks: classification, regression, clustering, association rule mining, and feature selection. It supports extensive data preprocessing options such as cleaning, normalization, and attribute transformation within a flexible pipeline, enhancing model performance on real-world data. One of WEKA’s hallmark features is its graphical user interface, which includes the Explorer, Experimenter, and KnowledgeFlow, enabling users to design, test, and visualize machine learning workflows interactively without requiring deep programming expertise. For advanced users, WEKA provides an API and scripting capabilities to integrate its algorithms into custom applications [35].
Technically, WEKA is implemented in Java, making it highly portable across operating systems and easy to extend with new algorithms. It processes input in the attribute-relation file format (ARFF) but also supports various data formats through database connectivity and file import/export. WEKA’s comprehensive collection of algorithms includes decision trees, support vector machines, neural networks, Bayesian classifiers, clustering methods like k-means, and association rule algorithms, making it a versatile toolset for empirical machine learning experimentation and deployment.
Mathematically, WEKA’s algorithms collectively represent a spectrum of machine learning methods, from statistical models and decision theory to heuristic and optimization-based techniques. The suite supports comparative evaluation approaches including cross-validation and receiver operating characteristic (ROC) analysis, facilitating rigorous empirical assessment of model performance—an essential pillar of machine learning research methodology.
WEKA’s impact is significant as it democratized access to machine learning techniques by lowering the barrier to entry and standardizing experimentation workflows. It remains a cornerstone in machine learning education and research, continuously updated by the University of Waikato community and an active global user base. Its open-source nature has stimulated wide adoption in academia and industry, fostering innovation, collaboration, and reproducibility in the machine learning field [36].

3.5. Support Vector Machines (SVM, 1995)

Support Vector Machines (SVM), a transformative method in machine learning, were introduced in their widely recognized form through the seminal 1995 paper by Corinna Cortes and Vladimir Vapnik at AT&T Bell Laboratories. Building on foundational work by Vapnik and Alexey Chervonenkis dating back to the 1960s and 1970s, SVM formalized a powerful approach to binary classification by finding the optimal separating hyperplane that maximizes the margin—the distance between the hyperplane and the nearest data points of each class, called support vectors. This concept of maximum-margin classification provides robust generalization on unseen data, improving predictive accuracy and reducing overfitting compared to many earlier algorithms [37].
Mathematically, given training data points { ( x i , y i ) } i = 1 n where x i R p and labels y i { 1 , + 1 } , SVM finds a hyperplane defined by w · x + b = 0 such that the margin 2 w is maximized under the constraints
y i ( w · x i + b ) 1 , i .
. The resulting convex optimization problem can be solved efficiently using quadratic programming. To handle non-linearly separable data, Cortes and Vapnik introduced the "soft margin" formulation, allowing some misclassification controlled by slack variables, balancing margin maximization and error minimization. Furthermore, the kernel trick—developed in continuation by Boser, Guyon, and Vapnik in 1992—enables SVMs to operate implicitly in high-dimensional feature spaces using kernel functions like polynomial or radial basis functions, thereby extending SVMs to complex, nonlinear classification tasks.
SVM’s rigorous roots in statistical learning theory give it strong theoretical guarantees on generalization and consistency, distinguishing it from heuristic-based classification methods of its time. Its algorithmic efficiency, versatility across domains, and theoretical elegance made it a core machine learning tool from the mid-1990s onward, with applications spanning image recognition, bioinformatics, text classification, and many others. The 1995 paper by Cortes and Vapnik is regarded as a breakthrough that not only provided a practically viable classification algorithm but also solidified the importance of large-margin classifiers in supervised learning.
Today, SVM remains a backbone method in classical machine learning, often contrasted with modern deep learning approaches but still valued for its interpretability, mathematical clarity, and performance on smaller datasets. Its influence continues in enhancing kernel methods, support vector regression, and structured prediction tasks, making it a foundational pillar in the evolution of AI and machine learning [38]

3.6. OpenCV (2000)

OpenCV (Open Source Computer Vision Library) is an influential open-source library for computer vision and image processing that originated as an Intel Research initiative in 1999 under the leadership of Gary Bradski. Its first public release occurred in 2000, aiming from the start to democratize computer vision by providing optimized, portable, and accessible software infrastructure for a broad array of vision tasks. This initiative sought to eliminate duplication of effort by offering a comprehensive and efficient collection of algorithms for real-time image and video analysis, benefiting researchers, developers, and commercial applications alike [39].
Developed primarily in C and C++ with bindings for Python, Java, and other languages, OpenCV has evolved into a cross-platform [40] library widely used in academia and industry for its breadth and efficiency. Its architecture includes modular components for image processing, feature detection, object recognition, camera calibration, machine learning, and later, integration of deep learning modules for neural network inference. OpenCV’s development ethos stresses performance optimization, with ongoing enhancements including multi-core processing, GPU acceleration, and support for new hardware platforms to meet the demands of modern applications such as autonomous vehicles, augmented reality, robotics, and medical imaging.
Mathematically, OpenCV encapsulates a wide array of methods from classical computer vision like edge detection, geometric transformations, and stereo vision to statistical and machine learning techniques, including clustering, support vector machines, and neural networks. It provides implementation for linear algebra, matrix operations, filtering, and advanced models used in image analysis, facilitating both foundational research and applied system development.
Over the years, OpenCV’s stewardship transitioned from Intel to Willow Garage and Itseez before returning under Intel’s portfolio via acquisition, with an active community and the non-profit OpenCV.org foundation now guiding its continual evolution. Its broad adoption, sustained development, and extensible architecture make OpenCV a cornerstone tool that has significantly influenced the growth and accessibility of computer vision technologies for more than two decades [41].

3.7. MATLAB AI Toolbox (2000s)

MATLAB, originally developed in the late 1970s by Cleve Moler and commercialized through MathWorks from the mid-1980s onwards, is a powerful numeric computing environment and programming language widely adopted in engineering, science, and applied mathematics. The MATLAB AI Toolbox, introduced progressively throughout the 2000s, significantly expanded MATLAB’s capabilities by providing a dedicated platform for developing and experimenting with machine learning, artificial intelligence, and control algorithms. These toolboxes integrated a rich variety of functions and apps supporting classification, regression, clustering, neural networks, reinforcement learning, and deep learning, along with tools for data preparation, visualization, algorithm tuning, and deployment [42].
Architecturally, the AI Toolbox leverages MATLAB’s matrix-based language to allow users to express ML and AI algorithms succinctly, combining high-level programming with efficient built-in computational routines. The toolbox supports both algorithmic prototyping and production-grade code generation, enabling seamless transitions from research experiments to implementation. MATLAB’s interactive environment with integrated plotting and coding facilitates fast iteration, debugging, and data exploration—making it particularly favored in academic research and industrial applications focused on control systems, autonomous vehicles, robotics, signal processing, and biomedical engineering.
Mathematically, MATLAB AI Toolbox algorithms encompass classical statistical models, optimization routines, neural network training via backpropagation [43], and advanced deep learning architectures . The platform provides robust support for linear and nonlinear system modeling, state estimation, and adaptive control theory, underpinned by extensive numerical libraries for matrix decomposition, eigenvalue computation, numerical integration, and optimization algorithms such as gradient descent, conjugate gradient, and quasi-Newton methods. Deep learning support integrates seamlessly with popular frameworks like TensorFlow and PyTorch, allowing MATLAB users to leverage and customize cutting-edge network architectures.
The MATLAB AI Toolbox has become a cornerstone in the AI research and development ecosystem due to its comprehensive functionality, ease of use, and powerful computational engine. It promotes reproducibility and collaboration across disciplines by combining rich algorithmic libraries with user-friendly interactive features. Continuously enhanced by MathWorks, the toolbox plays a critical role in advancing AI technologies, enabling engineers and scientists to develop new AI methods and apply them to complex real-world problems efficiently [44] .

3.8. Scikit-learn (2007–2010)

Scikit-learn (also known as sklearn), developed between 2007 and 2010, is a widely acclaimed open-source machine learning library for the Python programming language designed to provide accessible and efficient tools for data mining and data analysis. The project began as a Google Summer of Code initiative by French data scientist David Cournapeau in 2007, originally intended as a scientific toolkit extension to the SciPy ecosystem. In 2010, key contributors from the French Institute for Research in Computer Science and Automation (INRIA)—Fabian Pedregosa, Gaël Varoquaux, Alexandre Gramfort, and Vincent Michel—took over the development, leading to the release of the first public version on February 1, 2010.
Scikit-learn’s architecture is built around a consistent, user-friendly API that integrates a broad selection of supervised and unsupervised learning algorithms, including classification, regression, clustering, dimensionality reduction, and model selection. It builds on the foundational scientific computing libraries in Python, notably NumPy and SciPy, enabling efficient numerical computation and making it a natural choice for researchers and data scientists. The library also provides comprehensive documentation, concrete examples, and seamless interoperability with other Python data tools such as pandas and Matplotlib.
Mathematically, scikit-learn encompasses a diverse range of algorithms spanning linear models (like linear and logistic regression), support vector machines, ensemble methods (random forests, gradient boosting), clustering methods (k-means, DBSCAN), and manifold learning, among others. It supports parameter tuning, model validation through cross-validation, and evaluation metrics, facilitating rigorous experimental workflows in machine learning research and industry applications.
Scikit-learn’s impact has been significant in democratizing machine learning, making state-of-the-art algorithms easy to experiment with and deploy while fostering a large active community that continuously contributes improvements. Its design philosophy emphasizes simplicity, efficiency, and versatility, contributing to its widespread adoption across academic and commercial sectors, advancing both education and practical AI solutions worldwide.

4. Deep Learning & Neural Network Revolution (2010s)

4.1. Theano (2010)

Theano is an open-source Python library developed at the Montreal Institute for Learning Algorithms (MILA) at the Université de Montréal, first released publicly in 2007. It was designed as a powerful numerical computation framework that specializes in defining, optimizing, and efficiently evaluating mathematical expressions, particularly those involving multi-dimensional arrays, which are common in machine learning and deep learning models. Its core innovation lay in symbolic computation, where users define computational graphs symbolically, allowing Theano to apply sophisticated optimizations and compile efficient code targeted for CPU or GPU hardware, crucially enabling GPU-accelerated training of complex neural networks [45].
Developed under the leadership of renowned AI researcher Yoshua Bengio and his team, Theano introduced automated differentiation capabilities, simplifying the process of computing gradients necessary for training deep learning models using backpropagation. By abstracting the mathematical details and transparently leveraging hardware acceleration, Theano helped researchers explore and implement novel neural architectures and advanced learning algorithms with greater ease and speed than was previously possible.
Mathematically, Theano’s strength comes from representing computations as directed acyclic graphs with nodes representing mathematical operations and edges representing data dependencies, facilitating efficient symbolic differentiation through the chain rule and optimization of the computation graph before execution. This design allowed it to seamlessly integrate with the NumPy ecosystem while offering dynamic C code generation and extensive unit testing for numerical stability and reliability.
Though major development ceased in 2017 due to emerging competitors like TensorFlow and PyTorch, Theano’s influence endures—it served as the computational backend for popular deep learning libraries such as Keras and Lasagne, and its pioneering work in symbolic graph optimizations and automatic differentiation continue to underpin modern deep learning frameworks. The open-source community and projects like PyTensor (a fork and continuation) have maintained its legacy, ensuring its foundational ideas persist in advancing AI research and applications.
Overall, Theano was pivotal in accelerating the adoption and research of deep learning by combining mathematical rigor, computational efficiency, and GPU acceleration, marking a critical milestone in the evolution of AI tools

4.2. Caffe (2013)

Caffe, short for Convolutional Architecture for Fast Feature Embedding, is an open-source deep learning framework developed at the Berkeley Vision and Learning Center (BVLC) by Yangqing Jia in 2013 during his PhD at UC Berkeley. Designed with expressiveness, speed, and modularity as primary goals, Caffe quickly became a popular choice for researchers and practitioners working on convolutional neural networks (CNNs) and other deep learning models, especially in computer vision. The framework allows users to define, train, and deploy deep networks using configuration files without hard-coding, which encourages easy experimentation and innovation [46].
Caffe’s architecture is built in C++ with a Python interface, providing flexibility and performance, including seamless switching between CPU and GPU computations by setting a single flag. It supports a wide range of neural network components such as convolutional, fully connected, and recurrent layers, and integrates NVIDIA’s cuDNN library for GPU acceleration, enabling it to process over 60 million images per day on a single NVIDIA K40 GPU. Caffe was among the fastest convolutional network implementations available, making it highly suitable for both academic research experiments and industrial-scale deployments in vision, speech, and multimedia.
Mathematically, Caffe supports typical building blocks of deep learning: convolutions, pooling, activation functions, normalization, dropout, and backpropagation for gradient-based optimization. It leverages an efficient computation graph with layer-wise modularity, allowing various networks to be constructed by composing simple, reusable components. Its efficiency allowed state-of-the-art architectures like AlexNet and GoogleNet to be implemented and evaluated easily, accelerating breakthroughs in image classification and other domains [47].
The development of Caffe marked a significant milestone in deep learning research by emphasizing performance, modularity, and ease of use, fostering an extensive community of contributors and users. Although its direct development waned after 2018 with the advent of newer frameworks like TensorFlow and PyTorch, Caffe’s core ideas and implementations live on, influencing modern deep learning tools. It remains important historically for accelerating the adoption and advancement of CNNs in computer vision and AI research [48]

4.3. TensorFlow (2015)

TensorFlow is an open-source deep learning framework developed and released by the Google Brain team in 2015 [49]. It emerged as the successor to Google’s earlier proprietary system, DistBelief, and was designed to provide a flexible, scalable, and efficient platform for implementing machine learning and deep learning models. TensorFlow adopts a dataflow programming model where computations are expressed as stateful dataflow graphs consisting of nodes representing operations and edges representing multidimensional data arrays called tensors, from which the framework derives its name. Its architecture supports distributed computing and seamless deployment across CPUs, GPUs, and Google’s custom Tensor Processing Units (TPUs), enabling training and execution of complex neural networks on diverse hardware from desktops to large server clusters and mobile devices.
Initially released under the Apache 2.0 open-source license, TensorFlow aimed to accelerate AI research and democratize access to powerful machine learning tools. Since its launch, TensorFlow has become the most popular and widely adopted deep learning framework globally, thanks to its combination of computational efficiency, comprehensive ecosystem, and extensive community support. It includes modules for a range of AI tasks such as model training, evaluation, prediction, visualization, and deployment. TensorFlow’s Keras API integration, introduced from version 2.0 onwards, further simplified neural network programming by offering a high-level, user-friendly interface.
Mathematically, TensorFlow enables the construction of flexible computational graphs that represent complex tensor operations. These graphs support automatic differentiation, a critical component for efficient backpropagation used in training deep neural networks. The system’s parallel and distributed design leverages optimization techniques and low-level integration with hardware drivers to maximize throughput and reduce training time. Its layered architecture abstracts details of hardware acceleration and memory management, allowing researchers and developers to focus on model innovation rather than implementation complexity.
TensorFlow’s impact extends beyond academia and research labs into industry, powering applications from image recognition to natural language processing and autonomous systems. Its continuous development by Google and an active open-source community ensures it remains a cornerstone of AI advancement and deep learning innovation worldwide [50].

4.4. Keras (2015)

Keras, introduced in 2015 by François Chollet, a Google engineer, is an open-source, high-level neural network API designed to simplify the development and experimentation with deep learning models. Developed as part of the ONEIROS research project (Open-ended Neuro-Electronic Intelligent Robot Operating System), Keras was motivated by the desire to make deep learning more accessible, modular, and user-friendly, targeting rapid prototyping and easy experimentation in neural network design [51]. Unlike low-level frameworks that require detailed knowledge of tensor operations and computational graphs, Keras offers a clean, intuitive API centered around building models layer-by-layer, which greatly reduces the cognitive load on developers.
Initially, Keras was independent, capable of running on multiple backend engines including TensorFlow, Theano, and Microsoft Cognitive Toolkit (CNTK). Starting from version 2.4, it was integrated into TensorFlow as its official high-level API through the tf.keras module, leveraging TensorFlow’s computational efficiency while maintaining its signature simplicity. The modular design of Keras allows users to build complex neural network architectures from simple building blocks like layers, activation functions, optimizers, and loss functions, all customizable and extensible. Its design philosophy prioritizes ease of use without sacrificing flexibility, making it popular among researchers, students, and industry practitioners [52].
Mathematically, Keras encapsulates standard deep learning components such as dense layers (fully connected), convolutional layers, recurrent neural networks (LSTM, GRU), and allows automatic differentiation and backpropagation via its backend. It supports a range of optimization algorithms like stochastic gradient descent and Adam, and facilitates integration of custom loss functions or metrics. Keras’s abstraction thus enables seamless experimentation with complex architectures while handling the underlying tensor operations and graph optimizations.
Keras’s influence is profound as it lowered barriers to entry into deep learning research and development, accelerated experimentation cycles, and facilitated widespread adoption of neural networks. Its integration into TensorFlow and continued development positions Keras as a cornerstone tool in modern AI pipelines, beloved for balancing simplicity and power. François Chollet’s original vision to democratize deep learning technology continues to resonate as Keras scales to support new frameworks and hardware.

4.5. PyTorch (2016)

PyTorch is a dynamic, open-source deep learning framework developed by Facebook AI Research (FAIR) and first released in 2016. Created as a successor to the Lua-based Torch framework, PyTorch was designed to provide researchers and developers with a more intuitive, Pythonic environment for building neural networks. By adopting a dynamic computational graph architecture—often described as define-by-run—PyTorch enabled program execution to be flexible and adaptable, allowing models to be modified on the fly during training or inference, which significantly simplified experimentation and debugging compared to static graph frameworks like TensorFlow [53].
The development team at FAIR, led by Soumith Chintala among others, focused on usability and seamless integration with the Python data science ecosystem. PyTorch supports accelerated tensor computation on CPUs and GPUs, automatic differentiation via the autograd system, and a rich set of predefined neural network components. In 2018, PyTorch 1.0 was released, merging the research-oriented PyTorch with production-ready Caffe2, unifying flexibility and scalability to streamline deployment from prototype to industrial applications.
Mathematically, PyTorch enables efficient representation and optimization of neural networks through dynamic construction and traversal of computation graphs. It supports standard deep learning operations, including convolutions, recurrent layers, and various activation and loss functions, combined with optimization algorithms like stochastic gradient descent and Adam. The autograd engine automates gradient computation via reverse-mode differentiation, integral to backpropagation training.
PyTorch’s flexible interface, broad community support, and comprehensive tools have made it the framework of choice in AI research laboratories, academic settings, and increasingly in production environments. Its ecosystem now encompasses libraries for vision (TorchVision), natural language processing (TorchText), reinforcement learning, and more, while its adoption by major tech companies highlights its critical role in advancing AI innovation. The establishment of the PyTorch Foundation under the Linux Foundation in 2022 formalized its open governance, ensuring ongoing development guided by an industry consortium representing leading technology enterprises [54].
PyTorch’s prominence stems from blending the agility needed for research with the demands of production, making it a flagship deep learning platform that continues to evolve at the forefront of AI development.

4.6. CNTK (2016)

Microsoft Cognitive Toolkit (CNTK), originally unveiled in 2016, is an open-source deep learning framework developed by Microsoft primarily for scalable machine learning across multiple GPUs and distributed systems. The roots of CNTK trace to internal Microsoft research needs, especially for accelerating speech and language processing projects—including models behind the Cortana virtual assistant and Bing web ranking. CNTK was created with a focus on performance, efficiency, and flexibility, aiming to democratize robust AI tools by making them available to researchers and practitioners through open-source licensing. Initially released with a proprietary scripting language known as BrainScript, CNTK quickly evolved to offer high-level APIs for Python, C++, and later C#, broadening its accessibility [55].
The framework models neural networks as computational graphs, allowing for intuitive representation and manipulation of feed-forward, convolutional, and recurrent architectures. CNTK supports automatic differentiation (backpropagation) and includes optimized readers for efficient handling of sparse and dense data in machine learning tasks. Its design enables parallelization and distributed learning—even over thousands of GPUs—utilizing innovations such as 1-bit stochastic gradient descent for highly efficient cross-node communication.
Mathematically, CNTK enables state-of-the-art model training with rigorous implementation of deep learning constructs, from matrix and tensor operations to symbolic recurrent loops and multi-server parallelization. In addition to standard optimization algorithms, it supports advanced gradient propagation and memory sharing to maximize hardware utilization. A hallmark of CNTK is its seamless scalability, facilitating practical training of large neural models for image, speech, text, and time-series data.
CNTK found widespread use within Microsoft and externally, distinguishing itself by pioneering large-scale, efficient model training and offering strong integration with Windows and Linux environments. Although its prominence has declined in favor of frameworks like TensorFlow and PyTorch, CNTK’s technical innovations remain influential in large-scale machine learning infrastructure. Its last major release, version 2.7, supports ONNX interoperability and legacy AI projects, with many concepts and optimization strategies contributing to modern AI software engineering [56].

4.7. MXNet (2015)

MXNet, officially known as Apache MXNet, is a scalable, open-source deep learning framework that was first released in 2015 and widely adopted by Amazon Web Services (AWS) as its preferred deep learning library. MXNet originated as a joint research effort between the University of Washington and Carnegie Mellon University, with key contributions from Carlos Guestrin and collaborators. The framework quickly gained traction in the industry for its unique combination of flexibility, efficiency, and distributed computing capabilities, making it ideal for both academic research and enterprise-scale production environments [57].
MXNet’s architecture is distinguished by its hybrid programming model, seamlessly integrating both symbolic (static, declarative) and imperative (dynamic, Pythonic) approaches. Users can describe complex neural networks as computation graphs—benefiting from graph-level optimizations for memory and performance—or operate directly with tensor computations for maximum flexibility and debugging ease. A core innovation is the dynamic dependency scheduler that automatically parallelizes computation over CPUs and GPUs, and efficiently scales across multiple devices and nodes using a distributed parameter server for fast synchronization and data exchange.
Mathematically, MXNet supports building deep neural networks for supervised and unsupervised learning, including convolutional neural networks (CNNs), recurrent models (LSTMs, GRU), and advanced architectures. Its symbolic execution abstractions facilitate automatic differentiation and graph optimization for backpropagation, while imperative APIs offer intuitive manipulation of NDArray tensors. MXNet’s scalability allows near-linear compute performance increases as GPU and CPU resources are added, which has been a decisive factor in handling large datasets and sophisticated models in production-scale applications.
MXNet’s extensive language support—including Python, R, Scala, Julia, and more—and compatibility with major cloud platforms like AWS, Microsoft Azure, and edge devices, further fueled its adoption. In 2017, MXNet became an Apache Top-Level Project, recognized for its robust governance and active community contributions. Despite its declining development activity due to industry shifts toward frameworks like PyTorch and TensorFlow, MXNet’s legacy persists through its innovations in multi-language support, scalability, and cloud-native design [58].
Today, MXNet serves as an important case study in deep learning framework design, underpinning many of Amazon’s AI services and continuing to influence large-scale neural network research and deployment strategies.

4.8. DeepMind AlphaGo (2016)

DeepMind AlphaGo, developed by DeepMind (a subsidiary of Alphabet/Google), made history in 2016 as the first artificial intelligence system to defeat a reigning world champion in the ancient board game of Go—a feat long considered out of reach for machines due to Go’s astronomical complexity and intuitive nature. Go presents a vast search space—estimated at 1017010170 possible board states—rendering brute-force search or traditional rule-based AI approaches inadequate. AlphaGo’s innovation lay in integrating deep neural networks with advanced reinforcement learning and search techniques, bridging the gap between human-like intuition and rigorous calculation [59].
The AI’s core architecture combined two deep neural networks: the policy network, trained using both supervised learning on millions of human expert moves and subsequently refined with reinforcement learning through self-play, proposes promising moves; and the value network, trained to predict the winner from a given board position, guides the evaluation of Go states. These networks underpin a massively parallelized Monte Carlo Tree Search (MCTS) algorithm, which simulates future play sequences and strategically explores the most relevant lines of play, prioritizing actions the policy network deems most probable and evaluating states through the value network. The networks themselves are convolutional neural systems with millions of parameters, capable of automatic feature extraction from raw board representations.
AlphaGo’s training began by imitating human play from a large corpus of expert games, followed by intensive reinforcement learning wherein the system played countless games against versions of itself, refining its strategy far beyond human knowledge. The reinforcement learning step applies the policy-gradient approach, adjusting network weights to maximize board position values, guided by rewards based on game outcomes. Through this, AlphaGo developed powerful “intuition” for high-level play, even surprising expert players by inventing novel, creative moves.
AlphaGo’s triumph over Lee Sedol in 2016 inspired the AI world, marking a paradigm shift in what reinforcement learning and neural networks could achieve, especially when combined with powerful computational resources and innovative training protocols. Its legacy extends beyond Go: the architecture and algorithms pioneered by AlphaGo have influenced domains as diverse as robotics, protein folding (AlphaFold), and resource optimization, demonstrating the transformative potential of deep reinforcement learning in solving problems of enormous complexity and strategic depth [60].

4.9. OpenAI Gym (2016)

OpenAI Gym, introduced by OpenAI in April 2016, is a pivotal open-source toolkit specifically designed to advance the field of reinforcement learning (RL) by providing a standardized, extensible collection of simulation environments and benchmarks. The toolkit was developed to address two fundamental challenges in RL research: the lack of standardized environments for algorithm comparison and the need for reproducible benchmarks to objectively evaluate and refine RL algorithms. By unifying how environments are structured and interacted with, OpenAI Gym accelerated the pace and rigor of RL research, enabling rapid prototyping, fair benchmarking, and cross-comparison of algorithms in both academia and industry [61].
Technically, Gym offers a Python API and a modular design, making it easy to create, wrap, or extend environments ranging from classic control systems (e.g., CartPole, MountainCar) to Atari 2600 games, physics-based robotics simulations (using MuJoCo, Box2D, or PyBullet), and custom, user-built tasks. Each environment adheres to a simple agent-environment interface: an agent observes a state, takes an action, receives a reward and a new state, and determines when an episode ends. This abstraction standardizes reinforcement learning experimentation, while the expanding collection of built-in environments accommodates both discrete and continuous action spaces, stochastic dynamics, and varying levels of task complexity. Gym’s architecture supports seamless integration with popular deep learning frameworks like TensorFlow and PyTorch, allowing RL agents to leverage powerful neural networks for function approximation, policy learning, and value estimation.
OpenAI Gym also played a critical role in fostering an ecosystem around open RL research, inspiring the development of extensions like Gymnasium (its current community-maintained incarnation) and libraries for multi-agent RL, robotics, and benchmarking. Leading RL algorithms—from traditional Q-learning and SARSA to modern advancements in deep reinforcement learning—have been evaluated and reproduced using Gym’s environments, which contributed to the field’s rapid progress and reproducibility. Its influence extends to educational initiatives, practical applications in robotics, gaming, and AI modeling for complex sequential-decision problems [62].
In sum, OpenAI Gym’s innovation lies in its standardization, extensibility, and community-driven design, which together underpin the contemporary landscape of reinforcement learning research, education, and application development

4.10. H2O.ai (2015)

H2O.ai, founded in 2012, is an open-source artificial intelligence and machine learning company with a vision of democratizing AI for individuals and businesses. Its most recognized platform, H2O, and subsequent AutoML offerings emerged as early leaders in delivering powerful, accessible machine learning solutions suitable for practitioners with varying technical expertise. The core H2O platform, commonly called H2O-3, was developed to provide scalable, distributed machine learning—capable of handling massive datasets in memory and natively integrating with Big Data frameworks like Hadoop and Apache Spark. This foundation made H2O widely used in both academic research and industry, thanks to its extensibility and high computational performance [63].
The introduction of H2O AutoML first released in 2017—propelled the H2O platform into the era of automation. AutoML lowers the barrier to entry for building machine learning models by automating processes such as data preprocessing, feature engineering, model selection, hyperparameter tuning, and generation of stacked ensemble models. Users can build and deploy high-quality predictive models with minimal coding using intuitive APIs in R, Python, Java, and Scala, or the point-and-click H2O Flow web GUI. Under the hood, H2O AutoML efficiently conducts randomized model searches and ensemble learning to create a ranked leaderboard of models, balancing predictive accuracy with computational efficiency. Its distributed "H2O Cluster" architecture allows scaling across multi-node, multi-core environments, suitable for both on-premise and cloud deployments [64].
H2O.ai distinguishes itself through robust community participation and open-source ethos, along with features emphasizing interpretability, transparency, and regulatory compliance—a cornerstone for finance, healthcare, and high-stakes domains. Its Driverless AI product takes automation further, leveraging advanced techniques to build, optimize, and explain models automatically, making state-of-the-art AI more accessible and reliable.
With its blend of automation, scalability, and interpretability, H2O.ai has solidified its influence in the AI and data science landscape, empowering organizations of all types to leverage AutoML for fast, efficient, and explainable model building and deployment.

5. Generative & Pretrained Model Era (2017–2020)

5.1. Transformer Architecture (2017)

The Transformer architecture, introduced by Vaswani et al. in the landmark 2017 Google paper “Attention Is All You Need,” revolutionized the field of natural language processing, sequence modeling, and—ultimately—became the foundation for modern large language models (LLMs). Prior to Transformers, neural sequence modeling was dominated by recurrent neural networks (RNNs) and long short-term memory (LSTM) architectures, which processed input data in order, making parallelization difficult and struggling with long-range dependencies [65].
The breakthrough of Transformers was their reliance solely on attention mechanisms—specifically, self-attention—to process inputs in parallel and capture context dependencies across entire sequences, regardless of their position. The model consists of an encoder and a decoder built from stacks of identical layers, each featuring multi-head self-attention and feedforward neural networks. Self-attention computes attention scores for each element, allowing the model to dynamically focus on the most relevant parts of input sequences when deriving contextual representations. Multi-head attention enables the model to consider multiple representation subspaces simultaneously, dramatically enhancing its capacity to encode nuanced relationships within data [66].
Mathematically, for a sequence of input vectors X , self-attention computes output vectors by weighting the importance of every other element using attention scores derived from scaled dot products. This mechanism is expressed as:
Attention ( Q , K , V ) = softmax Q K T d k V
where Q , K , and V are the query, key, and value matrices derived from X , and d k denotes the dimensionality of the keys. This structure enables both efficient parallel computation and clear mathematical optimization, in contrast to the iterative nature of RNNs.
The impact of Transformer architecture has been enormous. It forms the backbone of all subsequent foundational models in NLP and beyond, such as Google’s BERT, OpenAI’s GPT series, and many others [67] . These models deliver unprecedented performance in understanding and generating human language, powering everything from search engines to chatbots, content summarizers, and coding assistants. The ability to scale Transformers—by increasing their depth, width, and training datasets—led directly to the explosion of LLMs, with models reaching billions of parameters and previously unattainable capabilities. Research since 2017 has extended these principles to computer vision, speech, and multimodal AI, confirming Transformers as the defining innovation of the current era of machine learning and artificial intelligence.

5.2. BERT (2018)

BERT (Bidirectional Encoder Representations from Transformers), released by Google in 2018, marked a major leap in contextual language understanding and has become the cornerstone for many modern natural language processing (NLP) applications. Developed by Jacob Devlin and colleagues, BERT introduced the idea of deep, bidirectional pre-training that considers both the left and right context of every word in a sentence at every layer—unlike previous models (such as word2vec, GloVe, or even unidirectional Transformer-based models like GPT) that process language in a single direction or with only shallow bidirectionality [68].
Built on the encoder segment of the Transformer architecture, BERT leverages self-attention to create contextual embeddings, enabling it to capture subtle meanings of words based on the entire sentence. The original BERT model was released in two standard sizes: BERT_BASE (12 layers, 12 attention heads, 110M parameters) and BERT_LARGE (24 layers, 16 attention heads, 340M parameters). Its bidirectional context modeling allows BERT to resolve ambiguous words with high accuracy, making it significantly more powerful for tasks like question answering, sentiment analysis, and coreference resolution.
BERT’s training involves two novel self-supervised objectives:
  • Masked Language Modeling (MLM): Randomly masks words in a sentence and trains the model to predict them using the surrounding context, fostering deep bidirectional understanding [69].
  • Next Sentence Prediction (NSP): Trains the model to predict whether a given sentence logically follows another, enhancing its ability to grasp sentence relationships and discourse structure.
After pre-training on massive corpora, BERT can be fine-tuned with just an additional output layer for virtually any NLP task, and achieves state-of-the-art results across a broad spectrum of benchmarks. Its open-source release sparked a surge of innovation, spawning a family of models (e.g., RoBERTa, ALBERT, DistilBERT) and fundamentally changing the standard paradigm for NLP: from training task-specific models from scratch to leveraging large, pre-trained “foundation” language models and adapting them for downstream tasks.
BERT’s adoption by Google Search in 2019 brought its impact to billions of users worldwide, dramatically improving search query understanding for over 70 languages. BERT’s introduction established the pre-train-then-fine-tune model as the dominant approach to language understanding and catalyzed the development of ever-larger and more powerful language models, ushering in the post-BERT era in AI [70].

5.3. GPT-2 (2019)

GPT-2 (Generative Pre-trained Transformer 2), released by OpenAI in February 2019, was a landmark breakthrough in generative language modeling. Building directly on its predecessor GPT-1, GPT-2 featured a dramatic leap in scale—with up to 1.5 billion parameters—and showcased the power of unsupervised learning on massive datasets. Trained on the WebText corpus (8 million web pages, 4 ~ 0GB), GPT-2 demonstrated the ability to produce multi-paragraph, coherent, and contextually relevant text with only a simple prompt, making it suitable for tasks such as story generation, translation, question answering, summarization, and even composing poetry [71].
Its architecture is a decoder-only Transformer—a stack of multi-head self-attention and feedforward layers—operating autoregressively (predicting the next word given preceding text). GPT-2’s design allows it to model long-range dependencies and generate fluid, adaptive prose. The model’s “chameleon-like” flexibility enabled strong zero-shot and few-shot learning: GPT-2 could perform a variety of tasks without any task-specific training data, simply by interpreting instructions given in natural language [72].
OpenAI opted for a staged release, initially limiting public access to smaller models before fully releasing the largest 1.5-billion-parameter version in November 2019. This cautious approach was due to concerns over misuse, such as generating convincing fake news, spam, or impersonation content—GPT-2 was one of the first AI models to raise broad public debate over the ethical risks of advanced text generation technology. Despite these concerns, GPT-2 rapidly became a research baseline, demonstrating that scaling up Transformers led to dramatic improvements in text generation quality and general-purpose language understanding [73].
GPT-2’s impact is immense: it inspired a wave of larger, more capable models (GPT-3, GPT-4), catalyzed the development of conversational AI, creative writing tools, and powerful text-based assistants, and underscored the paradigm of language model pre-training followed by task-specific adaptation. Its code and checkpoints remain widely used in research and industry, and its success marked the beginning of the large language model (LLM) revolution [74].

5.4. StyleGAN (2019)

StyleGAN, introduced by Nvidia researchers and made source-available in February 2019, is a groundbreaking generative adversarial network (GAN) architecture that set a new benchmark in ultra-realistic image synthesis—most notably for generating convincing portraits of non-existent human faces. Building upon previous GAN advancements, especially Nvidia’s Progressive GAN (ProGAN, 2017), StyleGAN’s innovation lies in its ability to provide fine-grained control over visual attributes (“styles”) at different levels of image abstraction, from the coarse (face shape, pose) to the fine (hair texture, freckles, wrinkles).
The core of StyleGAN architecture uses a mapping network that transforms a random latent vector into a set of “style” vectors [75]. These style vectors are injected at multiple layers of the generator, allowing for highly controllable synthesis of image features. This “adaptive instance normalization” based mechanism means users can smoothly interpolate features—mixing, blending, and manipulating aspects of generated faces with remarkable realism. Standard GAN training is retained, where a generator learns to create images and a discriminator learns to distinguish real from fake, iteratively improving the realism of outputs. [166]
StyleGAN’s technological leap became widely recognized through the viral website “This Person Does Not Exist,” which showcased the ability to generate endless, lifelike human faces with each refresh. Its impact reverberated through both creative and scientific domains: StyleGAN is used in art, game graphics, synthetic data generation, and education about media authenticity. Following the original release, Nvidia improved the method with StyleGAN2 (2020), which removed visual artifacts and enhanced image quality, and StyleGAN3 (2021), which solved “texture sticking” and delivered more alias-free generation—further refining the consistency of generated details.
Mathematically, StyleGAN’s generator employs convolutional neural layers where style vectors modulate normalization parameters. This enables controlled variation in features and “style mixing”—a capability earlier GANs lacked. The resultant images regularly surpass previous models in realism as measured by metrics such as the Fréchet Inception Distance (FID).
In summary, StyleGAN’s introduction opened new possibilities in visual synthesis, enabling ultra-realistic image generation, fine feature manipulation, and creative exploration. Its releases and iterative improvements remain critical milestones in the history of AI-generated imagery and machine creativity [76].

5.5. T5 (2019)

T5 (Text-to-Text Transfer Transformer), introduced by Google Research in 2019, is a transformative language model that unified how natural language processing (NLP) tasks are framed and solved. Unlike earlier models that treated tasks such as translation, summarization, and question answering with custom architectures or approaches, T5 proposed a “text-to-text” framework: every problem is cast as converting an input text string into an output text string, regardless of the underlying NLP task [77].
Built on the full encoder-decoder Transformer architecture, T5 processes input text using the encoder and generates output text with the decoder, leveraging the power of attention mechanisms throughout. The unifying principle is to prepend tasks with special instruction-like text prefixes (e.g., “translate English to German: That is good”) and train the model [165]end-to-end on a diverse range of tasks using the same architecture and loss function. This provides a consistent interface and allows multitask learning, where the same model can be fine-tuned or prompted for a wide array of downstream applications—including translation, summarization, classification, question answering, and more [78].
For pre-training, T5 uses a large, high-quality dataset called the Colossal Clean Crawled Corpus (C4), containing hundreds of gigabytes of web-scraped English text. The self-supervised learning objective is “span corruption,” where random spans of text are replaced with sentinel tokens and the model learns to reconstruct the missing content, enabling deep contextual and compositional language understanding. T5’s vocabulary is built using SentencePiece tokenization, allowing coverage of multiple languages and efficient handling of rare or out-of-vocabulary words [79].
By recasting every NLP task as a text transformation, T5 not only simplified the training and deployment process but also delivered state-of-the-art performance across a diverse array of language benchmarks. Its influence is seen in subsequent developments in “instruction tuning” and general-purpose, instruction-following LLMs, which still trace their roots to T5’s unified approach [80].
T5 represents a major evolution in NLP thinking, demonstrating the power of casting all language problems into a single, flexible text-to-text mold while leveraging large-scale transfer learning [81].

5.6. FastAI (2018)

FastAI, launched in 2018 by Jeremy Howard and Sylvain Gugger, is an open-source deep learning library built atop PyTorch that is designed to make state-of-the-art machine learning accessible to beginners and experts alike. FastAI abstracts and automates many of the complexities of deep learning, enabling rapid experimentation, intuitive workflows, and best practices by default, while still allowing for full access to underlying PyTorch capabilities [82].
The library introduces a layered API:
  • High-level: Functions for common deep learning tasks (vision, text, tabular, time series, collaborative filtering) that minimize code and domain-specific knowledge requirements.
  • Mid-level/Low-level: Modular building blocks that let advanced users customize model architectures, training strategies, and data preprocessing pipelines.
Core FastAI concepts include:
  • DataBlock/DataLoaders: Cleanly structured tools for scalable, flexible data preprocessing and loading.
  • Learner: Encapsulates a complete model training pipeline—bringing together data, network architecture, training/evaluation logic, and reporting.
  • Built-in Best Practices: Automated data augmentation, mixed precision training, transfer learning integration, and state-of-the-art optimizers simplify robust model development.
Mathematically, FastAI leverages PyTorch for tensor operations, differentiation, and GPU acceleration. Its API extends PyTorch’s flexibility with powerful abstractions that enable researchers to prototype, train, and deploy models with less boilerplate and more focus on innovation.
FastAI’s vibrant ecosystem is complemented by comprehensive courses, documentation, and a large community, which has contributed to making cutting-edge AI techniques approachable for practitioners, educators, and researchers across the world. It has been instrumental in democratizing deep learning education and practice, accelerating the adoption of PyTorch and modern deep learning best practices in both industry and academia [83].

5.7. AllenNLP (2018)

AllenNLP, launched in 2018 by the Allen Institute for Artificial Intelligence (AI2), is an open-source research library built on top of PyTorch that is dedicated to advancing natural language processing (NLP) research and applications. AllenNLP addresses the common challenges faced by NLP researchers—such as reproducibility, extensibility, and ease of experimentation—by providing reusable building blocks, modular data pipelines, and configuration-driven experiment management [84].
Key architectural features of AllenNLP include:
  • PyTorch Foundation: Leveraging PyTorch’s dynamic computation graphs, enabling flexible model design and intuitive debugging.
  • Modular Components: Reusable modules for tokenization, data reading, embedding, encoding, and pre/post-processing, which allow rapid prototyping and efficient pipeline construction.
  • Declarative Configurations: JSON or Python-based experiment configurations, making it easy to define, reproduce, and share experimental workflows, models, and hyperparameters.
  • Reference Implementations: High-quality models for a variety of NLP tasks, such as semantic role labeling, textual entailment, question answering, and named entity recognition, help users benchmark and extend cutting-edge research methods.
  • Flexible Data API: A “Field” and “Instance” abstraction allows unified and efficient handling of diverse NLP data structures, such as sequences, spans, and trees, with automatic sorting, batching, and padding.
AllenNLP is especially valued in the NLP research community for streamlining new model development and fostering reproducibility. It also comes with tools for visualization, evaluation, and integration of pre-trained models. Its APIs and pipelines have made it a popular choice for both academic research and real-world NLP productization [85].
Applications range from text classification, semantic parsing, and coreference resolution to information extraction and knowledge graph construction. AllenNLP continues to be actively developed and widely adopted, facilitating rapid advances in NLP research and supporting open science through a thriving community and extensive documentation.

5.8. Hugging Face Transformers (2019)

Hugging Face Transformers, launched in 2019, rapidly became the most influential open-source platform and library for accessing, sharing, and deploying pre-trained models in modern natural language processing (NLP). Initially, Hugging Face was focused on chatbot development, but its founders soon recognized the profound potential—and the community’s need—for a unified hub that would make cutting-edge transformer models (beginning with BERT, GPT-2, and others) easily accessible to all practitioners and researchers[86] . The result was the Transformers library—a consistent API and repository that supports PyTorch and TensorFlow, and enables seamless downloading, fine-tuning, and deployment of models for tasks such as text classification, question answering, translation, summarization, and more[87].
This platform’s breakthrough was to “democratize” access to the most powerful models [164] resulting from the “transformer revolution,” which had previously been confined to specialized labs or required significant engineering expertise to reproduce. In the Hugging Face ecosystem, anyone can load a model with a line of code, experiment interactively, and share improvements or new models via the hosted Model Hub. It soon evolved from supporting a handful of models to hosting thousands—including nearly all major transformer-based architectures like RoBERTa, T5, DistilBERT, XLNet, and later GPT-3, BLOOM [88] , and large multimodal models. Alongside the core library, Hugging Face released companion tools for tokenization, dataset management, and evaluation, making it not only a toolkit for inference but a full-stack environment for research, production, and benchmarking.
The impact on NLP can hardly be overstated. By radically lowering barriers to entry, Hugging Face enabled rapid experimentation, broad collaboration, and the sharing of reproducible results—accelerating progress on major benchmarks, downstream applications, and even in languages and domains with less well-funded research. Its open-source ethos cultivated a global community of contributors and users ranging from machine learning engineers and academic labs to major tech companies and startups. As a result, Transformer-based methods became standard practice for a huge swath of industry and academia [89].
Moreover, Hugging Face Transformers now plays a critical role in shaping the landscape of generative AI, powering conversational agents, language generators, document classifiers, and bioinformatics solutions. It bridges research and production, providing high-level interfaces suitable for non-experts and deeply customizable options for advanced users. Its effective blend of technical excellence, user experience, and community-driven development has solidified its status as the “GitHub for machine learning models”—a foundational resource in today’s AI ecosystem, and a major force in the continuing evolution of large language models and intelligent systems [90].

6. Multimodal & Generative AI Explosion (2020–2023)

6.1. GPT-3 (2020)

GPT-3 (Generative Pre-trained Transformer 3), released by OpenAI in 2020 [91] , marked a major milestone in the evolution of generative AI. With a staggering 175 billion parameters, GPT-3 dwarfed all previous language models—being over a hundred times larger than GPT-2—and signaled a new era of artificial intelligence powered by scale and transfer learning. Like its predecessors, GPT-3 is a decoder-only Transformer model, built on deep neural network layers that leverage self-attention to analyze and synthesize input text. It was trained on a vast corpus—over 45 terabytes of diverse text from Common Crawl, Wikipedia, books, and scientific articles—allowing it to attain fluid, humanlike language abilities.
The most remarkable advance with GPT-3 is its zero-shot and few-shot learning capabilities. With just minimal or no task-specific fine-tuning, GPT-3 can generate text, answer questions, translate languages, summarize, solve arithmetic problems, write computer code, and adapt to complex prompts—all by interpreting context and examples given at inference time. This “prompt engineering” approach allows users to unlock new behaviors from a single model without retraining, a radical leap compared to previous generation systems. The model’s context window of 2048 tokens enables understanding and retention of lengthy passages, contributing to coherent multi-paragraph output [92].
GPT-3’s emergence resulted in widespread impact across industries. It powers chatbots, creative writing assistants, educational tools, content automation systems, and is the core technology behind OpenAI’s later product ChatGPT (initially based on GPT-3.5, later upgraded to GPT-4). Its performance on standardized NLP benchmarks set new records, and its general-purpose text generation was so convincing that many human evaluators struggled to distinguish its writing from authentic human prose [93]. The scale of GPT-3 also raised concerns about ethical risks, misinformation, and the social responsibility of deploying powerful language models, prompting debates that continue as models grow even larger and more capable.
Licensing and API access to GPT-3 reflected both excitement and caution; Microsoft secured exclusive rights to the underlying model, while others access it via OpenAI’s cloud API. In the wake of GPT-3, the field experienced a surge of competitive innovation—spurring subsequent large models from Google, Meta, and open initiatives like EleutherAI [94].

6.2. CLIP (2021)

CLIP (Contrastive Language–Image Pretraining), introduced by OpenAI in 2021, represents a key milestone in multimodal AI by fundamentally bridging the world of images and natural language. Unlike classical computer vision models, which require extensive supervised training for each specific task, CLIP learns to associate images and text in a unified representation space through a process called contrastive learning. The model was trained on an unprecedented scale—400 million image–text pairs sourced from the internet—allowing it to develop a broad and general understanding of visual concepts as expressed in natural language [95] .
The technical architecture of CLIP consists of two separate neural networks: a vision encoder (typically a Vision Transformer or ResNet) and a text encoder (usually based on a Transformer). Both encoders map their respective inputs—an image and a piece of text—to high-dimensional embeddings in a shared feature space. During training, CLIP maximizes the cosine similarity between embeddings for matching image–caption pairs and minimizes it for mismatched pairs. The outcome is a model that can, given an image and a set of candidate text descriptions (or vice versa), identify which text best matches the image—even for objects or situations it has never seen before [96] .
What makes CLIP extraordinary is its "zero-shot" capability: without any further task-specific training, it can classify images, retrieve relevant images given text queries, generate image captions, and more, simply by leveraging the rich relationships embedded in its multimodal representation. For instance, CLIP can instantly label previously unseen images by choosing from a list of natural language prompts, far surpassing the flexibility of models reliant on tightly controlled class labels or datasets. Its approach to generalization also inspired the architecture of numerous AI systems powering text-to-image synthesis, content moderation, and robust semantic search [97] .
CLIP’s release was also transformative at the ecosystem level. It became an essential building block for more advanced multimodal models, including guidance systems for generative models like DALL-E and Stable Diffusion. In industry and research, CLIP powers applications ranging from search engines and recommendation systems to digital art, content filtering, and the investigation of neural network interpretability (such as "multimodal neurons"). Its public release has enabled wide experimentation and rapid progress in building truly general-purpose, flexible AI capable of connecting language and vision [98] .

6.3. DALL·E (2021)

DALL·E, unveiled by OpenAI in January 2021, is a pioneering deep learning model designed to generate novel images from text prompts—a capability that captured the world’s attention and redefined the potential of generative AI. Building on the success of large language models like GPT-3, DALL·E employs a Transformer-based architecture and leverages concepts from autoregressive models and variational autoencoders (VAEs). It was trained on hundreds of millions of image–text pairings collected from the internet, allowing the model to synthesize strikingly creative and highly coherent visuals based solely on detailed natural language descriptions [99].
The technical breakthrough of DALL·E lies in representing both text and images as discrete tokens. During training, a discrete VAE compresses each image into a lower-dimensional grid of tokens, while the Transformer learns to model the joint distribution of text and image tokens as a single sequence [100]. When given a prompt, DALL·E generates the sequence of image tokens that, when decoded by the VAE, produces an image matching the semantic content of the text. This means a prompt like “an armchair in the shape of an avocado” will yield an entirely new image, blending previously unseen concepts and styles in photorealistic or artistic ways.
One of DALL·E’s most significant impacts is its seemingly limitless generative creativity and compositional reasoning. The model can merge disparate ideas, invent objects, adapt styles, or transform concepts much like an imaginative artist responding to verbal instructions. DALL·E’s “zero-shot” ability—generating meaningful imagery for prompts it was never explicitly trained on—quickly found applications in art, design, education, and synthetic data generation [101].
Following the original DALL·E, OpenAI released DALL·E 2, which improved image quality and capability using advances in diffusion models guided by CLIP, OpenAI’s multimodal text–image embedding network. These innovations established a new paradigm for generative models: text-to-image (and now text-to-video and text-to-3D) synthesis that democratizes visual creativity and speeds up workflows in media, design, and scientific visualization. Both the original DALL·E and its successors have become benchmarks, inspiring further development in open-source generative models and autonomous creative systems worldwide [102]

6.4. Stable Diffusion (2022)

Stable Diffusion, released in August 2022 by Stability AI in partnership with CompVis and Runway ML, transformed the generative AI landscape by making photorealistic image synthesis open-source and accessible to everyone—even those without enterprise-level hardware or budgets. Unlike previous proprietary models like DALL·E or Midjourney, Stable Diffusion’s source code and model weights were openly licensed, allowing anyone to download, run, and modify the system for personal, academic, or creative use. This choice democratized AI-based image generation, fueling an explosion of community-driven innovation and applications across art, media, research, and industry [103] .
Technically, Stable Diffusion is a latent text-to-image diffusion model. It works by first compressing images into a lower-dimensional latent space using an autoencoder, then employing a neural network to iteratively reverse random noise back into a coherent picture, guided by a textual prompt. The choice of working in latent space—as opposed to the full pixel space—allows for faster image generation and reduces computational requirements, enabling high-quality results even on consumer GPUs. For understanding prompts, Stable Diffusion incorporates OpenAI’s CLIP text encoder, which maps natural language instructions to vector representations closely aligned with the visual domain, ensuring that generated images faithfully reflect the user’s input [104] .
Open-source accessibility led to massive adoption and rapid evolution. Creators and developers built custom interfaces, extensions, and plug-ins, while artists and designers gained powerful new tools for ideation, prototyping, and artistic expression. The flexibility to modify and fine-tune models spawned countless niche variants and tailored solutions, expanding the possibilities for commercial, educational, and creative uses. Stable Diffusion also prompted important conversations around ethics, copyright, and the future of creative work, particularly as human artists contended with new forms of digital art generation and content authenticity [105] .
Culturally and technologically, Stable Diffusion is credited with “removing the doors from their hinges”—ushering in a new era where AI image generation became a grassroots phenomenon rather than a privilege for major tech companies. With billions of images created and a vibrant community fueling continued improvement, Stable Diffusion remains a backbone of generative art, synthetic media, and open research, pressuring proprietary competitors to open up and spurring advances throughout the AI field [106] .

6.5. Midjourney (2022)

Midjourney, launched in early 2022, quickly established itself as one of the leading platforms for AI-driven art generation, harnessing the power of text-to-image synthesis to empower creators of all kinds. Founded by David Holz (also known for Leap Motion), Midjourney is an independent research lab whose eponymous software enables users to craft striking, imaginative visuals from simple text prompts. Unlike previous art generators focused primarily on photorealism or technical demonstration, Midjourney is distinguished by its painterly aesthetics, creatively blending real-world styles, artistic influences, and fantastic elements. Its model excels in fantasy scenes, stylized environments, and expressive character portraits a favorite among concept artists, designers, illustrators, and hobbyists [107].
Access to Midjourney is innovative: the platform operates primarily as a bot on Discord, where users type prompts and receive images instantly, sparking communal sharing, feedback, and collaborative exploration. This workflow makes AI art creation interactive and social, driving the rapid growth of a global user base that has surpassed 16 million by late 2023. Midjourney offers continual updates—releasing improved algorithms every few months, each lifting artistic quality, coherence, prompt accuracy, and stylistic variety. By 2024, its web interface expanded accessibility, bringing AI visual generation beyond Discord’s audience [108].
Under the hood, while Midjourney is speculated to use principles similar to latent diffusion models (as in Stable Diffusion), its proprietary technology and blend of artistic tuning set it apart. The system leverages high-quality image–text pairing, learning complex associations between language and visual style, and allowing highly customized creations based on user instructions. The result is a new form of visual ideation, letting professionals and enthusiasts prototype commercial art, create book illustrations, design product concepts, or simply explore creative possibilities, often in minutes [109].
Midjourney’s impact stretches far beyond its technical achievements. Its accessible, community-centered approach to AI-generated art has helped disrupt traditional stock photography, lowered creative barriers, and brought sophisticated art generation to the masses. The platform has powered magazine covers, children’s books, concept art, and even winning entries in digital art competitions. Users praise its ability to speed up brainstorming, explore unfamiliar styles, and develop visual narratives with unprecedented ease [110].

6.6. Whisper (2022)

Whisper, released by OpenAI in September 2022, is an open-source automatic speech recognition (ASR) model that set a new standard for transcribing and translating audio across a diverse range of languages, accents, and audio environments. Its development was motivated by the need for robust, accurate, and scalable speech-to-text systems not just for English but for over 100 languages worldwide—addressing limitations of earlier models that struggled with non-standard accents, domain-specific jargon, and noisy backgrounds [111].
At the heart of Whisper’s approach is a weakly-supervised training pipeline and a sequence-to-sequence encoder-decoder Transformer architecture. The model takes as input a log-Mel spectrogram—a time-frequency representation—of audio split into 30-second chunks, transforming it into a sequence of latent vectors that capture temporal, spectral [112], and semantic information. The decoder then generates transcriptions token by token, with the same model capable of language identification, transcription in the original language, translation into English, and phrase-level timestamps . This multifunctional design was achieved by training on an exceptionally large and diverse dataset of 680,000 hours of supervised audio-text pairs sourced from the web, which enabled the model to learn a wide variety of accents, recording conditions, and technical vocabularies.
Whisper’s release as open source offered unprecedented benefits for developers, researchers, and practitioners. It provided not only high-accuracy English transcription but also robust multilingual transcription and speech-to-English translation without any extra fine-tuning. The model’s versatility and ease of use sparked rapid adoption, enabling better accessibility solutions, improved voice interfaces, podcast and video transcription, language study tools, and facilitating research into more complex audio and speech applications [113].
Whisper’s ability to generalize across languages, accents, and environments has made it a foundational technology in the global push for inclusive, AI-powered speech and language tools. Moreover, its transparent release set a benchmark for openness in AI development, supporting further innovation in speech recognition technology and its integration into both commercial and open-source projects [114].

6.7. ChatGPT (2022)

ChatGPT, released by OpenAI in November 2022, represents one of the most remarkable leaps forward for conversational artificial intelligence. Built initially on GPT-3.5 and later enhanced by GPT-4, ChatGPT marked the first time a large language model was fine-tuned and deployed specifically for natural dialogue with broad, everyday usability. Its launch was a watershed moment: millions adopted ChatGPT within days, integrating it into personal and professional routines for Q&A, drafting content, brainstorming, tutoring, technical assistance, and creative writing. The system’s ability to provide coherent, contextually relevant, and human-like responses—often with nuanced understanding and personality—demonstrated the practical viability of large-scale conversational AI for general use [115].
ChatGPT’s core advances derive from two pivotal techniques. First, the conversational interface was built to handle multi-turn interactions, enabling users to clarify, elaborate, and correct threads of dialogue in natural language. Second, the model was fine-tuned via reinforcement learning from human feedback (RLHF), aligning responses with human preferences, improving safety, and minimizing bias and harmful outputs. This combination made ChatGPT surprisingly capable at not just answering questions, but engaging in back-and-forth exchanges, reasoning through ambiguity, admitting mistakes, and playfully challenging incorrect premises [116].
ChatGPT’s underlying models evolved rapidly: starting with GPT-3.5, which offered strong general-domain language abilities, then moving to GPT-4 in March 2023, whose larger scale, improved alignment, and multimodal capabilities pushed comprehension and creativity even further. By 2024—2025, versions with even broader context windows, enhanced reasoning, and tool integration were released, enabling the system to use web search, interpret images, code, and analyze user data. Throughout, model architecture remained based on the Transformer paradigm, with progressive improvements in alignment, factuality, steerability, and reasoning depth [117].
The impact of ChatGPT is profound and ongoing. It has been widely adopted in customer service, education, content creation, programming, healthcare, entertainment, and more—changing workflows and user expectations about interacting with computers. Its ease of use—running in browsers, apps, and APIs—helped it reach billions and inspire a new wave of competition among tech companies [118].
Perhaps even more importantly, ChatGPT popularized prompt engineering and conversational design as skills, enabled researchers to probe both the powers and limitations of LLMs, and raised important ethical, legal, and social questions about synthetic dialogue, misinformation, and AI-assisted work [119].

6.8. BLOOM (2022)

BLOOM (BigScience Large Open-science Open-access Multilingual Language Model), released in July 2022, is a landmark achievement in the movement toward open, collaborative, and transparent AI research. Developed over a year-long workshop by the BigScience collaboration—a vast consortium coordinated by Hugging Face and comprising more than a thousand researchers from around the world—BLOOM is one of the largest large language models ever made publicly accessible. With 176 billion parameters, BLOOM stands alongside proprietary models like GPT-3 and PaLM but distinguishes itself by distributing both its model weights and training data under permissive open licenses, allowing for unrestricted research, adaptation, and scrutiny worldwide [120].
The architecture of BLOOM is a transformer-based autoregressive language model, trained to generate text and code across 46 natural languages and 13 programming languages. Its multilingual proficiency was a crucial focus, making it the most widely inclusive open-source LLM of its time and a powerful tool for countering the English-centric bias seen in earlier AI models. To achieve this, BLOOM was trained on the diverse, transparent ROOTS corpus, containing nearly 1.6TB of text carefully sourced, documented, and cleaned to reflect not just the web but also underrepresented communities and linguistic groups [121].
Transparency extended far beyond the code and weights—the whole process, including engineering decisions, training logs, and ethical deliberations, was carried out in public view on the Jean Zay supercomputer in France. This pioneering approach allowed researchers everywhere to follow, audit, and contribute at every step, setting a new gold standard for open science and reproducibility in the AI field [122].
BLOOM’s contributions extend beyond its technical prowess. By making such a model openly available, it democratized large-scale AI, enabling communities, universities, and smaller innovators—especially those from developing regions—to participate in frontier machine learning research without the heavy restrictions or costs associated with proprietary models. This has fostered not only novel applications and research but also greater global dialogue around bias, governance, multilingual access, and responsible AI development [123].

6.9. LLaMA (2023)

LLaMA (Large Language Model Meta AI), released by Meta (formerly Facebook) in early 2023, represents a major advancement in accessible large language model research. Developed as a suite of transformer-based models with parameter sizes ranging from 7 billion to 65 billion for the first generation (and up to 70 billion in subsequent versions), LLaMA was designed to rival performance benchmarks set by proprietary LLMs like GPT-3 and PaLM, but with a focus on efficiency, transparency, and open availability to the global research community [124].
The initial LLaMA models were provided under a research license, and while they were intended for non-commercial use and distribution to approved academic applicants, their weights were rapidly disseminated online, catalyzing a burst of experimentation and fine-tuning by independent researchers worldwide. This “leak” incident, while controversial, contributed to the immediate formation of an active ecosystem: academics and developers built instruction-following variants like Alpaca (Stanford), Vicuna (LMSYS), and Koala, showing that state-of-the-art conversational AI could be achieved for relatively low cost and infrastructure by fine-tuning LLaMA foundations [125].
Meta’s follow-up, LLaMA 2 (July 2023), further revolutionized access to advanced LLMs. LLaMA 2 models—released in 7B, 13B, and 70B parameter sizes—offered not just foundational model weights but also chat-optimized, instruction-following variants with enhanced safety fine-tuning. Notably, LLaMA 2’s more permissive license enabled commercial use under certain conditions and included wide platform integration with partners like Microsoft and Hugging Face, dramatically expanding its adoption in industry and research. The release of Code LLaMA, specialized for code generation and programming, further deepened its impact on technical and developer communities[126].
Technically, LLaMA’s success is often attributed to highly efficient model design, extensive multilingual and public dataset pretraining, and rigorous evaluation on diverse benchmarks. Its training datasets were constructed to optimize both language understanding and generation while making models compact enough for fine-tuning and inference on standard hardware, democratizing LLM research and customization worldwide.
LLaMA’s open availability has catalyzed a global movement of “open foundation” model development. Researchers and organizations now routinely create specialized, localized, and safety-aligned derivatives, pushing the boundaries of large language models while encouraging transparency, reproducibility, and broader participation in generative AI. At the same time, the LLaMA family has raised important questions about responsible model sharing and open science, as the ease of adaptation brings both societal benefits and risks [127].

6.10. Gemini (Bard) (2023)

Gemini, unveiled by Google DeepMind in December 2023, is Google’s most advanced multimodal AI model to date—purpose-built to natively understand, generate, and reason across text, images, audio, video, and code. Gemini was designed from the ground up as a multimodal system, meaning that, unlike previous models which often combined separately trained modules, Gemini’s architecture was jointly trained on diverse data types. This unified approach enables the model to seamlessly process complex, interleaved sequences of language and visuals, handle elaborate reasoning tasks, and provide highly contextual, accurate responses to a broad array of queries involving multiple modalities [128].
The Gemini family offers several model sizes—Ultra, Pro, and Nano—optimized for different use cases, from high-performance research to on-device AI for smartphones. Notably, Gemini Ultra has set new records on over 30 major academic and industry benchmarks, including being the first model to surpass human expert performance on the Massive Multitask Language Understanding (MMLU) exam. Gemini excels at a wide range of tasks, such as reading and interpreting complex diagrams, infographics, and scientific documents; analyzing natural images and video; audio transcription and understanding; sophisticated mathematical and logical reasoning; and robust code generation spanning multiple programming languages [129].
Gemini’s multimodal strength offers unique advantages in solving challenging questions that blend visual, linguistic, and logical information, making it highly effective for scientific discovery, education, content creation, and enterprise analytics. The system’s architecture directly supports the ingestion of text, images, audio waveforms, and video frames, allowing for nuanced analysis and synthesis that bridges gaps between previously siloed data formats [130].
Gemini also powers Google’s Bard chatbot (now rebranded as Gemini across many Google services), bringing enhanced generative AI to Search, Assistant, Android devices, and developer platforms like Google AI Studio and Vertex AI. Its natively multimodal reasoning, broad multilingual support, and ability to extract insights from massive datasets are driving innovation across consumer and enterprise applications [131].
Critically, Gemini isn’t just a technical achievement—it marks the culmination of Google’s efforts to bring together DeepMind and Google Brain, leveraging global collaboration and resources to set new standards for open, scalable, and responsible AI development. As Gemini continues to evolve—with version updates such as Gemini 1.5 and on-device deployment via Pixel phones—it is poised to define the next wave of general-purpose, highly capable, and accessible artificial intelligence [132].

6.11. Claude (2023)

Claude, released by Anthropic in March 2023, is a family of large language models engineered to push the state of conversational AI with a strong emphasis on safety, ethical alignment, and user control. Developed by a team of former OpenAI researchers, Claude’s defining innovation is its use of “Constitutional AI”—a training approach that combines Reinforcement Learning from Human Feedback (RLHF) with a set of guiding principles or "constitution." This constitution, drawn in part from documents like the UN Universal Declaration of Human Rights, directs the model to behave harmlessly and helpfully, relying not only on human feedback but on its own self-critiques and revisions. In practice, this method allows Claude to generate, review, and improve its own outputs while aligning with clear ethical standards, minimizing risks of bias, misinformation, and harmful responses[133].
Anthropic continually expanded Claude’s capabilities, releasing successive versions that increased the maximum context window—up to 200,000 tokens, or roughly 500 pages of material. This enabled users to upload and process lengthy documents, and perform advanced summarization, document analysis, and text-based research tasks well beyond prior systems. With Claude 2 and beyond, the models could read, interpret, and assist with tasks using a diverse range of inputs, including PDFs and complex workflows.
Safety and reliability are central to Claude’s design. The model demonstrates lower rates of hallucination and harmful output, and in recent updates, incorporates mechanisms capable of ending conversations in extreme cases of persistently abusive user interactions—reflecting Anthropic’s commitment to both model and user “welfare.” These capabilities, along with steerable tone, persona, and user feedback responsiveness, make Claude a uniquely user-friendly system for business, education, programming, creative writing, and data analysis [134].
Anthropic’s approach has spurred significant debate on the balance between ethical safeguards and usability, as some critics claim the model’s refusal to answer certain benign requests constitutes an “alignment tax.” Nonetheless, the importance of transparency, privacy, and verifiable ethical alignment places Claude at the leading edge of responsible and trustworthy AI development [135].

7. The Current Multimodal & Agentic Era (2024–2025)

7.1. GPT-4 / GPT-4o (2024)

GPT-4 and its successor GPT-4o (the "o" stands for "omni"), released by OpenAI in 2024, are state-of-the-art, natively multimodal large language models that mark a major leap in AI’s ability to understand and generate not just text, but images and audio as well. Unlike earlier models that relied on separate subsystems stitched together for different input types, GPT-4o is built as a unified neural network. This design allows it to process text, images, and audio in any combination as both input and output—enabling highly fluid, human-like interactions that span language, vision, and sound [136].
The capabilities of GPT-4 and especially GPT-4o go far beyond traditional chatbot functionality. For example, GPT-4o can read and describe photographs, interpret graphs or handwritten notes, answer questions about images, and even process live video feeds or screen recordings. In voice mode, it engages in real-time, multi-turn voice conversations at natural speeds, outperforming many previous models in speech recognition, translation, tone, and sentiment understanding. The system can not only recognize and transcribe audio, but respond with synthesized speech, even singing or mimicking emotional cues. This deep, native multimodality is a breakthrough for building interactive AI assistants that can see, listen, speak, and even analyze the real world as fluidly as a human [137].
GPT-4o’s performance extends further. It demonstrates state-of-the-art results on vision, audio, and multilingual benchmarks, outpacing major competitors on a range of tasks. OpenAI rolled out the model initially with support for text and vision and progressively expanded to enable complex audio and video generation. The model is highly efficient compared to its predecessors, enabling faster and cheaper deployment for both developers and consumers. Variants like GPT-4o mini—smaller and leaner, but outperforming previous flagship models—make advanced multimodal AI accessible for a broad range of devices and applications, from cloud APIs to mobile phones [138].
In practice, GPT-4o opens up new use cases—math tutoring with spoken explanations and hand-drawn diagrams, real-time translation in voice calls, multimodal creative work (combining writing, image creation, and music), and seamless accessibility tools for visually or hearing impaired users. Its development is a foundational shift toward AI systems that can meaningfully and naturally interact across the full spectrum of human communication channels [139].
OpenAI’s releases have catalyzed innovation throughout the industry, inspiring rapid advances and new research in general-purpose, multimodal, and interactive artificial intelligence. As multimodal systems like GPT-4o become integrated into everyday life, they are transforming not only how we interact with technology, but how we understand and bridge the boundaries between text, images, sound, and human expression [140].

7.2. Sora (2024)

Sora, launched by OpenAI in February 2024, is a breakthrough text-to-video generative AI model that enables users to create realistic, detailed videos simply by describing them in natural language. Building on the advances of prior text-to-image models like DALL·E, Sora harnesses advanced diffusion and transformer architectures to generate up to one minute of high-quality video per prompt. Its release marked a pivotal moment in generative media, allowing artists, filmmakers, educators, marketers, and everyday users to produce animated content from imagination alone—whether that’s simulating natural landscapes, crafting surreal animation, or rendering cinematic scenes [141].
Technically, Sora innovates by combining the strengths of diffusion models (for rich, low-level visual texture generation) with transformers (for global compositional layout and logical reasoning across video frames). It processes videos as sequences of "patches," akin to how tokens represent words in language models, maintaining object constancy and smooth motion over time. Sora also integrates automatic recaptioning: before generating, GPT reinterprets and expands the user’s prompt to add necessary detail, boosting fidelity and capturing intended nuance. The model was trained on an internet-scale dataset of image and video pairs, enabling broad generalization to a wide variety of subjects, genres, and visual storytelling styles [142].
Safety and copyright features are integral to Sora’s design. All videos generated are watermarked to distinguish AI-created media from authentic footage, helping counter potential misuse—such as the creation of fake historical clips or misleading visual content. OpenAI introduced Sora in a phased manner, starting with “red teams” and creative professionals for adversarial testing and feedback, then expanding availability through ChatGPT Plus and Pro subscriptions, and rolling out integration for platforms and mobile devices [143].
Sora’s impact on the creative industry is profound. It opens new avenues for prototyping, ideation, rapid storyboarding, educational content creation, and democratizes access to high-end video generation for audiences that could never afford large animation teams or expensive CGI. As of 2025, Sora has spawned a wave of third-party tools, plugin extensions, and ongoing open research into enhanced motion coherence, unbiased video synthesis, and broader support for diverse visual styles. With continued innovation—including the unveiling of Sora 2 and further accessibility features—OpenAI’s video generation has set a new standard for multimodal AI creativity in the digital age [144].

7.3. Gemini 1.5 (2024)

Gemini 1.5, introduced by Google DeepMind in early 2024, is a next-generation multimodal AI model renowned for its unprecedented ability to process and reason over extremely long contexts—including millions of tokens of text, images, audio, video, and code. Building on the success of the original Gemini and its predecessors, Gemini 1.5 sets a new benchmark by enabling users to upload, analyze, and interact with entire books, codebases, multi-hour podcasts, movies, and massive multi-document datasets in a single prompt—unlocking capabilities well beyond previous large language models [145].
The model is natively multimodal, meaning it was trained to handle a mix of modalities from the outset. This allows it to traverse, summarize, and derive insights from content such as scanned documents, annotated screenshots, medical imaging, video clips, and audio records—sometimes even answering detailed questions about specific scenes or extracting information from hand-drawn sketches. In qualitative demonstrations, Gemini 1.5 is able to locate specific portions of sprawling novels, translate new languages from grammar references alone, and pinpoint crucial code segments for debugging across large repositories [146].
Gemini 1.5 comes in several variants, including Gemini 1.5 Pro and Gemini 1.5 Flash—each tailored for different speed and quality trade-offs. With a context window scaling up to 1 million tokens (and experimental support for even longer sequences), Gemini 1.5 dramatically expands the practical limits of AI-powered analytics, making it possible for individuals and organizations to query and reason over data sources previously deemed too large or complex [147].
Real-world applications are changing rapidly as a result. Enterprise users leverage Gemini 1.5 for deep analysis of regulatory filings, legal contracts, patient medical histories, and thousands of hours of sensor data. Developers use it to search large codebases and even repair them automatically. Educators and scientists deploy long-context input for curriculum analysis, linguistic research, and multimedia content creation, while creative professionals blend images, transcripts, and video to produce meaningful, data-driven stories.
The ability of Gemini 1.5 to seamlessly integrate and retain massive multisource inputs—combined with state-of-the-art recall and reasoning—represents a leap forward not only for multimodal understanding, but also for building more personal, insightful, and comprehensive AI applications across industries. Google’s iterative releases and open documentation further enable rapid exploration of new use cases, setting the pace for industry adoption and innovation in long-context artificial intelligence [148].

7.4. Mistral & Mixtral (2024)

Mistral and Mixtral, released in 2024 by Paris-based Mistral AI, represent a major advance in open-source large language models (LLMs) by offering impressive performance, compact size, and highly efficient outputs that compete directly with Meta’s LLaMA family and sometimes even closed models like OpenAI’s GPT-3.5. Mistral 7B is a 7.3-billion-parameter model optimized for speed, resource efficiency, and strong benchmark results. It stands out for outperforming LLaMA 2 13B on most tasks—especially English-language and code-related benchmarks—while running on standard hardware, making it a popular choice for both research and practical deployments. Key features such as Grouped-query attention (GQA) and Sliding Window Attention (SWA) enable fast inference and extended context handling, with all weights released under a permissive Apache 2.0 license for easy local or cloud use [149].
Mixtral is Mistral AI’s series of Sparse Mixture of Experts (SMoE) models, such as Mixtral 8x7B and 8x22B. These models introduce a clever architecture where inference uses only a subset of available parameters per input token—enabling large aggregate model sizes (up to 141B parameters) with actual runtime speed and resource usage closer to smaller models. Mixtral models offer wider context windows (up to 64,000 tokens or more), very fast throughput, and robust multilingual capabilities, supporting languages like English, French, German, Italian, Spanish, and more. Performance-wise, Mixtral outstrips LLaMA 2 70B and shows competitive results against GPT-3.5 in common benchmarks, particularly excelling in code generation and complex instruction following. The open weights, high cost efficiency, and ease of fine-tuning have driven widespread adoption among developers looking for scalable, customizable, and safe AI solutions without commercial licensing barriers [150].
Mistral AI has also released proprietary variants for business use, but it’s the open models that have democratized powerful LLM deployment and experimentation, making high-quality large language models feasible for SMEs, academic labs, and grassroots communities. Their transparent release strategy, multilingual focus, and strong handling of long sequences make Mistral and Mixtral key drivers in the open-source AI ecosystem, frequently leading the leaderboard for best open LLM performance.
In summary, Mistral and Mixtral are lightweight, high-performing open models that continue to set standards for efficiency and accessibility, empowering broad communities to harness advanced AI for language, coding, and reasoning at scale—with a commitment to both open science and practical application [151].

7.5. Claude 3 (2024)

Claude 3, released by Anthropic in March 2024, is a flagship family of large language models that set new standards in enhanced reasoning, multimodal understanding, and user-centric control. The Claude 3 lineup includes three models—Haiku (optimized for speed and cost), Sonnet (balancing capability and efficiency), and Opus (pushing boundaries for complex reasoning and creativity). Building on Anthropic’s earlier commitment to Constitutional AI and reinforcement learning from human feedback (RLHF), Claude 3 models showcase dramatic improvements in following complex instructions, analyzing and summarizing massive documents, interpreting images and diagrams, and solving intricate math, coding, and logic tasks [152] .
For the first time, Claude gained native multimodal abilities, able to process both text and images within the same prompt, making it effective for visual analysis—such as extracting information from charts or understanding meaning in technical diagrams. The Opus model in particular came with an industry-leading context window of 200,000 tokens (over 500 pages of text), and experimental features expanding that window up to 1 million tokens for select use cases. This allowed users to work with entire novels, extensive legal contracts, and large codebases in a single session—enabling deep, integrated research and analysis [153].
Claude 3 quickly gained attention for outperforming major rivals—including OpenAI’s GPT-4—on key benchmarks in graduate-level reasoning, factual accuracy, coding, and knowledge-intensive workflows. Later in 2024, the Claude 3.5 Sonnet variant further strengthened its lead in code generation, chart interpretation, multistep workflow comprehension, and image-to-text extraction, and introduced the Artifacts feature, enabling real-time code testing and SVG/web rendering within the chat interface [154].
Beyond accuracy, one of Claude 3’s most distinctive advances is its hybrid dual-mode reasoning, realized in versions like Claude 3.7 Sonnet. This allows users to choose between rapid responses for simple questions and deep, step-by-step reasoning for more complex problems, making the model highly flexible for both quick searches and detailed analyses. Anthropic’s iterative upgrades, including the ability for Claude to control desktop environments and automate multi-application workflows, point to a future where large language models serve as fully agentic digital coworkers.
Claude 3’s release marked a turning point for safe, reliable, and powerful AI applications in business, education, research, and creative industries—offering users control, transparency, and the ability to work fluidly across long, complex, and multimodal information streams [155].

7.6. DeepSeek (2025)

DeepSeek, introduced in early 2025 by a Hangzhou-based Chinese tech company, is a series of open-source, multimodal AI models engineered to deliver high-quality generative reasoning at a fraction of the computational cost and financial overhead typical of competitive LLMs like OpenAI’s GPT-4o or Meta’s LLaMA 3.1. The release of DeepSeek’s R1 model sent shockwaves through the industry, earning it a reputation as an “AI revolution” or “Sputnik moment” for its ability to rival or exceed the outputs of established leaders while slashing training costs (with estimates for DeepSeek V3 at just 6 m i l l i o n , c o m p a r e d t o a r o u n d 100 million for GPT-4) [156].
Technologically, DeepSeek’s models leverage a combination of Mixture of Experts (MoE) architectures, reinforcement learning (RL), and clever engineering optimizations that drive down both hardware and energy requirements. These advances make DeepSeek models smaller, more efficient, and remarkably pragmatic for developers and organizations looking to deploy robust multimodal AI—handling text, images, and, in some cases, audio with high fluency. The models are designed with open weights (freely available under an MIT license), transparent technical documentation, and increased accessibility for research and downstream customization, even in the face of restricted access to advanced Western GPUs [157].
In late January 2025, DeepSeek-R1 made headlines for achieving performance near that of closed models like OpenAI’s GPT-o1 while remaining entirely open for academic and commercial examination. Researchers have praised DeepSeek’s combination of efficiency, versatility, and openness—helping foster a true global “AI price war” as major Chinese tech giants followed suit, cutting the costs of access to their own models and accelerating AI affordability worldwide.
DeepSeek’s further developments, such as DeepSeek-GRM (Generative Reward Modeling) and self-principled critique tuning, have enhanced the model’s ability to perform advanced reasoning and self-evaluation, further narrowing the gap with leading proprietary systems. The company’s openness also extends to collaborations, as with Tsinghua University, showcasing a commitment to transparent progress and continual release of updated, efficient, and high-performing models for theorem proving, math, and general reasoning tasks [158].
In sum, DeepSeek represents a breakthrough in democratizing advanced AI: fast, low-cost, efficacious, and openly accessible, rapidly reshaping global competition and setting new benchmarks for efficient multimodal and generative AI research and applications.

7.7. Runway Gen-2 (2024–2025)

Runway Gen-2, unveiled in mid-2023 and refined through 2024–2025, is a revolutionary video synthesis model empowering AI-driven image-to-video and text-to-video generation for creators, marketers, educators, and filmmakers. Developed by Runway AI, Gen-2 builds on the foundation laid by Gen-1, which introduced video-to-video transformations, but moves further by enabling users to craft entirely new videos from scratch using simple text prompts or static images—no cameras or traditional filming required [159].
The technology behind Gen-2 employs multimodal latent diffusion, enabling the model to interpolate between visual frames for temporal coherence, maintain stylistic consistency, and generate motion that harmonizes with the input image or narrative described in text. The model supports a variety of modes:
  • Text-to-Video: Generate original videos using descriptive natural language.
  • Image-to-Video: Animate a given image, bringing static scenes and objects to motion.
  • Text + Image-to-Video: Combine textual instructions and images for nuanced, controllable video output.
  • Stylization and Render: Transfer the style of a provided image or prompt to video frames, or turn untextured renders into realistic scenes.
  • Storyboard: Convert a sequence of mockups into fully animated video narratives.
Gen-2’s capabilities also allow for targeted visual control—masking, customizing character trajectories, and applying consistent rendering styles across all frames. This makes Gen-2 a valuable tool for rapid prototyping, commercial ad production, educational content creation, and social media engagement. Its user-friendly web platform, alongside API integration, has democratized advanced video creation, letting non-experts achieve professional results without expensive CGI or editing teams [160].
In terms of industry and cultural impact, Gen-2 has helped drive the AI video market’s exponential growth, reducing production time and costs by as much as 75–90% and giving creative professionals new power to iterate, test, and engage audiences through personalized, innovative video content. As a growing ecosystem, Runway continues to evolve its generative video models—launching Gen-3 and Gen-4 for longer, higher fidelity, and more stylistically coherent video output, and collaborating with entertainment companies for bespoke production workflows [161].
Altogether, Runway Gen-2 marks a new era in AI-assisted multimedia, providing the foundation for accessible, efficient, and expressive visual storytelling powered entirely by generative intelligence.

7.8. Perplexity AI (2025)

Perplexity AI, by 2025, has evolved into a leading search and reasoning assistant, redefining how individuals and teams access knowledge, perform research, and make critical decisions. Unlike traditional search engines that simply return lists of web links, Perplexity integrates state-of-the-art large language models (such as GPT-4.1, Claude 4 Sonnet, and others) with real-time web retrieval and a transparent citation system. This fusion—often described as “generative retrieval”—means users receive direct, conversational answers grounded in up-to-date sources, with every statement linked to authoritative references for fact-checking, academic integrity, and professional reliability [162].
At its core, Perplexity uses a hybrid pipeline combining natural language processing (to interpret user queries and context), semantic search (to retrieve the most relevant documents from the live web and academic indices), and retrieval-augmented generation (RAG) to synthesize responses from multiple sources. For complex or research-heavy queries, its “Deep Research” feature performs multi-hop reasoning—evaluating, cross-verifying, and weaving together information from dozens (or hundreds) of documents automatically. This allows Perplexity to generate structured meta-analyses, literature reviews, or in-depth business/comparative reports within minutes, all fully citation-backed and with clear transparency into source credibility [163].
Key innovations include:
  • Direct citations and clickable sources on every answer, supporting instant verification.
  • Multi-model switching, so users can refine queries with different LLMs for varied perspectives.
  • Contextual memory for follow-up Q&A within a single conversational session.
  • Advanced filtering and custom search parameters (especially via API) for enterprise, research, or developer integrations.
  • Rapid trend analysis, market insight, academic meta-reviews, and technical troubleshooting—all in natural language.
Perplexity’s impact has been profound: it has empowered millions of researchers, students, knowledge workers, and business analysts with fast, reliable access to both broad and deep knowledge. Its transparent approach has also spurred a wave of “answer engine” innovation among major tech firms. The platform is widely used as a co-pilot for research, coding, policy, and decision-making workflows—lowering the bar for expertise and transforming everyday search into a dynamic, dialogue-based process that supports critical thinking, productivity, and accuracy.

References

  1. (2006) — Newell and Simon’s Logic Theorist: Historical Background and Impact on Cognitive Modeling. [CrossRef]
  2. (1964) — Experiments with a theorem-utilizing program. [CrossRef]
  3. (2022) — What is Mathematics and What Should it Be https://arxiv.org/pdf/1704.05560. [CrossRef]
  4. (2008) — Principia Mathematica and the Development of Automated Theorem Proving. [CrossRef]
  5. (2006) — Newell and Simon’s Logic Theorist: Historical Background and Impact on Cognitive Modeling. [CrossRef]
  6. (2023) — Role of tectonic stress and topography on repeated lateral dikes: application to the 1975-1984 Krafla and 2023-2025 Svartsengi rifting episodes in Iceland. https://link.springer.com/content/pdf/10.1007/s00146-023-01793-z.pdf. [CrossRef]
  7. (2024) — Representing Rule-based Chatbots with Transformers https://arxiv.org/pdf/2407.10949. [CrossRef]
  8. (2024) — ELIZA Reinterpreted: The world’s first chatbot was not intended as a chatbot at all https://arxiv.org/pdf/2406.17650. [CrossRef]
  9. (2023) — Role of tectonic stress and topography on repeated lateral dikes: application to the 1975-1984 Krafla and 2023-2025 Svartsengi rifting episodes in Iceland. https://link.springer.com/content/pdf/10.1007/s00146-023-01793-z.pdf. [CrossRef]
  10. (1988) — SHRDLU, Procedures, Mini-World. [CrossRef]
  11. (2005) — SHRDLU. [CrossRef]
  12. (2019) — Survey on frontiers of language and robotics https://www.tandfonline.com/doi/pdf/10.1080/01691864.2019.1632223?needAccess=true. [CrossRef]
  13. (1970) — Applications of “Artificial Intelligence” for Chemical Inference, VI. Approach to a General Method of Interpreting Low Resolution Mass Spectra with a Computer. [CrossRef]
  14. (1974) — Inference of molecular structure. [CrossRef]
  15. (2025) — Challenges for artificial cognitive systems https://arxiv.org/pdf/2505.20339. [CrossRef]
  16. (1977) — Production Rules as a Representation for a Knowledge-Based Consultation Program. [CrossRef]
  17. (1974) — A rule-based computer program for advising physicians regarding antimicrobial therapy selection https://dl.acm.org/doi/pdf/10.1145/1408800.1408906.
  18. (2018) — Studying microbial functionality within the gut ecosystem by systems biology https://genesandnutrition.biomedcentral.com/track/pdf/10.1186/s12263-018-0594-6. [CrossRef]
  19. (2024) — Logic Programming with PROLOG. [CrossRef]
  20. (2018) — WAM for everyone: a virtual machine for logic programming. [CrossRef]
  21. (2022) — Fifty Years of Prolog and Beyond https://arxiv.org/pdf/2201.10816. [CrossRef]
  22. (1978) — History of LISP. [CrossRef]
  23. (1980) — LISP - notes on its past and future https://dl.acm.org/doi/pdf/10.1145/800087.802782. [CrossRef]
  24. (1978) — History of LISP https://dl.acm.org/doi/pdf/10.1145/960118.808387. [CrossRef]
  25. (2022) — Introduction to Soar https://arxiv.org/pdf/2205.03854. [CrossRef]
  26. (1986) — Soar—A General Problem-Solving Architecture. [CrossRef]
  27. (2022) — Introduction to Soar https://arxiv.org/pdf/2205.03854. [CrossRef]
  28. (1986) — Soar—A General Problem-Solving Architecture. [CrossRef]
  29. (2024) — Role of tectonic stress and topography on repeated lateral dikes: application to the 1975-1984 Krafla and 2023-2025 Svartsengi rifting episodes in Iceland. https://link.springer.com/content/pdf/10.1007/s44267-024-00065-8.pdf. [CrossRef]
  30. (2025) — Evidence of sensory error threshold in triggering locomotor adaptations in humans. [CrossRef]
  31. (2024) — Role of tectonic stress and topography on repeated lateral dikes: application to the 1975-1984 Krafla and 2023-2025 Svartsengi rifting episodes in Iceland. https://link.springer.com/content/pdf/10.1007/s00348-024-03814-z.pdf. [CrossRef]
  32. (1986) — Learning representations by back-propagating errors. [CrossRef]
  33. (1986) — Learning representations by back-propagating errors. [CrossRef]
  34. (2016) — Introducing Machine Learning Concepts with WEKA https://hdl.handle.net/10289/13170. [CrossRef]
  35. (2016) — Introducing Machine Learning Concepts with WEKA https://hdl.handle.net/10289/13170. [CrossRef]
  36. (2025) — Bayesian Workflow for Generative Modeling in Computational Psychiatry. [CrossRef]
  37. (2011) — Soviet Archaeological Expedition as a Research Object. [CrossRef]
  38. (2020) — Robust Predictive Models in Clinical Data—Random Forest and Support Vector Machines https://link.springer.com/content/pdf/10.1007%2F978-3-030-47994-7_13.pdf. [CrossRef]
  39. (2024) — Image Processing Using OpenCV. [CrossRef]
  40. (2023) — A Comprehensive Review of YOLO Architectures in Computer Vision: From YOLOv1 to YOLOv8 and YOLO-NAS https://arxiv.org/pdf/2304.00501. [CrossRef]
  41. (2022) — Object Detection in 20 Years: A Survey https://arxiv.org/pdf/1905.05055. [CrossRef]
  42. (2020) — Machine Learning in Python: Main Developments and Technology Trends in Data Science, Machine Learning, and Artificial Intelligence https://www.mdpi.com/2078-2489/11/4/193/pdf?version=1587379966. [CrossRef]
  43. (2022) — Neural Networks and Learning Algorithms in MATLAB. [CrossRef]
  44. (2023) — Quantifying the Benefit of Artificial Intelligence for Scientific Research https://arxiv.org/pdf/2304.10578. [CrossRef]
  45. (2024) — Implementing a Hierarchical Deep Learning Approach for Simulating Multilevel Auction Data. [CrossRef]
  46. (2022) — Caffe con Troll: Shallow Ideas to Speed Up Deep Learning https://arxiv.org/pdf/1504.04343. [CrossRef]
  47. (2014) — Caffe. [CrossRef]
  48. (2022) — Comparative Study of Deep Learning Software Frameworks https://arxiv.org/pdf/1511.06435. [CrossRef]
  49. (2022) — TensorFlow: A system for large-scale machine learning https://arxiv.org/pdf/1605.08695. [CrossRef]
  50. (2023) — Artificial Intelligence Index Report 2023 https://arxiv.org/pdf/2310.03715. [CrossRef]
  51. (2017) — kerasR: R Interface to the Keras Deep Learning Library https://joss.theoj.org/papers/10.21105/joss.00296.pdf. [CrossRef]
  52. (2021) — Role of tectonic stress and topography on repeated lateral dikes: application to the 1975-1984 Krafla and 2023-2025 Svartsengi rifting episodes in Iceland. https://link.springer.com/content/pdf/10.1007/s10586-021-03240-4.pdf. [CrossRef]
  53. (2024) — Practical machine learning with PyTorch https://jose.theoj.org/papers/10.21105/jose.00239.pdf. [CrossRef]
  54. (2025) — PyG 2.0: Scalable Learning on Real World Graphs https://arxiv.org/pdf/2507.16991. [CrossRef]
  55. (2022) — A detailed comparative study of open source deep learning frameworks https://arxiv.org/pdf/1903.00102. [CrossRef]
  56. (2022) — Benchmarking State-of-the-Art Deep Learning Software Tools https://arxiv.org/pdf/1608.07249. [CrossRef]
  57. (2022) — MXNet: A Flexible and Efficient Machine Learning Library for Heterogeneous Distributed Systems https://arxiv.org/pdf/1512.01274. [CrossRef]
  58. (2021) — Role of tectonic stress and topography on repeated lateral dikes: application to the 1975-1984 Krafla and 2023-2025 Svartsengi rifting episodes in Iceland. https://link.springer.com/content/pdf/10.1007/s10586-021-03240-4.pdf. [CrossRef]
  59. (2016) — Mastering the game of Go with deep neural networks and tree search. [CrossRef]
  60. (2016) — Mastering the game of Go with deep neural networks and tree search. [CrossRef]
  61. (2022) — Benchmarking Deep Reinforcement Learning for Continuous Control https://arxiv.org/pdf/1604.06778. [CrossRef]
  62. (2024) — Open RL Benchmark: Comprehensive Tracked Experiments for Reinforcement Learning https://arxiv.org/pdf/2402.03046. [CrossRef]
  63. (2023) — h2oGPT: Democratizing Large Language Models https://arxiv.org/pdf/2306.08161. [CrossRef]
  64. (2024) — Role of tectonic stress and topography on repeated lateral dikes: application to the 1975-1984 Krafla and 2023-2025 Svartsengi rifting episodes in Iceland. https://link.springer.com/content/pdf/10.1007/s10462-024-10726-1.pdf. [CrossRef]
  65. (2022) — Attention Is All You Need https://arxiv.org/pdf/1706.03762. [CrossRef]
  66. (2023) — Introduction to Transformers: an NLP Perspective https://arxiv.org/pdf/2311.17633. [CrossRef]
  67. (2023) — Introduction to Transformers: an NLP Perspective https://arxiv.org/pdf/2311.17633. [CrossRef]
  68. (2025) — Research and Implementation of Text Classification Based on BERT Model https://dl.acm.org/doi/pdf/10.1145/3746709.3746721. [CrossRef]
  69. (2021) — Encoder-Attention-Based Automatic Term Recognition (EA-ATR). [CrossRef]
  70. (2024) — BERT: A Paradigm Shift in Natural Language Processing. [CrossRef]
  71. (2022) — Language Models are Few-Shot Learners https://arxiv.org/pdf/2005.14165.
  72. (2020) — GPT-3: Its Nature, Scope, Limits, and Consequences https://link.springer.com/content/pdf/10.1007/s11023-020-09548-1.pdf. [CrossRef]
  73. (2025) — Role of tectonic stress and topography on repeated lateral dikes: application to the 1975-1984 Krafla and 2023-2025 Svartsengi rifting episodes in Iceland. https://link.springer.com/content/pdf/10.1007/s43621-025-00815-8.pdf. [CrossRef]
  74. (2022) — Language Models are Few-Shot Learners https://arxiv.org/pdf/2005.14165.
  75. (2022) — StyleGAN-XL: Scaling StyleGAN to Large Diverse Datasets https://arxiv.org/pdf/2202.00273. [CrossRef]
  76. (2022) — State-of-the-Art in the Architecture, Methods and Applications of StyleGAN https://arxiv.org/pdf/2202.14020. [CrossRef]
  77. (2022) — Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer https://arxiv.org/pdf/1910.10683. [CrossRef]
  78. (2022) — Attention Is All You Need https://arxiv.org/pdf/1706.03762. [CrossRef]
  79. (2025) — A Comprehensive Overview of Large Language Models. [CrossRef]
  80. (2022) — Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer https://arxiv.org/pdf/1910.10683. [CrossRef]
  81. (2025) — Foundations of Large Language Models https://arxiv.org/pdf/2501.09223. [CrossRef]
  82. (2024) — Role of tectonic stress and topography on repeated lateral dikes: application to the 1975-1984 Krafla and 2023-2025 Svartsengi rifting episodes in Iceland. https://link.springer.com/content/pdf/10.1007/s12599-024-00851-0.pdf. [CrossRef]
  83. (2023) — Artificial Intelligence Index Report 2023 https://arxiv.org/pdf/2310.03715. [CrossRef]
  84. (2018) — AllenNLP: A Deep Semantic Natural Language Processing Platform https://www.aclweb.org/anthology/W18-2501.pdf. [CrossRef]
  85. (2023) — Catwalk: A Unified Language Model Evaluation Framework for Many Datasets https://arxiv.org/pdf/2312.10253. [CrossRef]
  86. (2022) — HuggingFace’s Transformers: State-of-the-art Natural Language Processing https://arxiv.org/pdf/1910.03771. [CrossRef]
  87. (2023) — Transformer models: an introduction and catalog https://arxiv.org/pdf/2302.07730. [CrossRef]
  88. (2024) — Large language models (LLMs): survey, technical frameworks, and future challenges. [CrossRef]
  89. (2023) — To Build Our Future, We Must Know Our Past: Contextualizing Paradigm Shifts in Natural Language Processing https://arxiv.org/pdf/2310.07715. [CrossRef]
  90. (2024) — Role of tectonic stress and topography on repeated lateral dikes: application to the 1975-1984 Krafla and 2023-2025 Svartsengi rifting episodes in Iceland. https://link.springer.com/content/pdf/10.1007/s12599-024-00851-0.pdf. [CrossRef]
  91. (2020) — GPT-3: Its Nature, Scope, Limits, and Consequences https://link.springer.com/content/pdf/10.1007/s11023-020-09548-1.pdf. [CrossRef]
  92. (2022) — Language Models are Few-Shot Learners https://arxiv.org/pdf/2005.14165.
  93. (2023) — Sparks of Artificial General Intelligence: Early experiments with GPT-4 https://arxiv.org/pdf/2303.12712. [CrossRef]
  94. (2024) — Large Language Models: A Survey https://arxiv.org/pdf/2402.06196. [CrossRef]
  95. (2022) — Reproducible scaling laws for contrastive language-image learning https://arxiv.org/pdf/2212.07143. [CrossRef]
  96. (2022) — Learning Transferable Visual Models From Natural Language Supervision https://arxiv.org/pdf/2103.00020. [CrossRef]
  97. (2023) — Demystifying CLIP Data https://arxiv.org/pdf/2309.16671. [CrossRef]
  98. (2023) — CLIP in Medical Imaging: A Survey https://arxiv.org/pdf/2312.07353. [CrossRef]
  99. (2022) — Learning Transferable Visual Models From Natural Language Supervision https://arxiv.org/pdf/2103.00020. [CrossRef]
  100. (2022) — Zero-Shot Text-to-Image Generation https://arxiv.org/pdf/2102.12092. [CrossRef]
  101. (2022) — Large-scale Text-to-Image Generation Models for Visual Artists’ Creative Works https://arxiv.org/pdf/2210.08477. [CrossRef]
  102. (2023) — Text-to-image Diffusion Models in Generative AI: A Survey https://arxiv.org/pdf/2303.07909. [CrossRef]
  103. (2023) — SDXL: Improving Latent Diffusion Models for High-Resolution Image Synthesis https://arxiv.org/pdf/2307.01952. [CrossRef]
  104. (2022) — Learning Transferable Visual Models From Natural Language Supervision https://arxiv.org/pdf/2103.00020. [CrossRef]
  105. (2023) — Art and the science of generative AI: A deeper dive https://arxiv.org/pdf/2306.04141. [CrossRef]
  106. (2025) — Role of tectonic stress and topography on repeated lateral dikes: application to the 1975-1984 Krafla and 2023-2025 Svartsengi rifting episodes in Iceland. https://link.springer.com/content/pdf/10.1007/s10462-025-11110-3.pdf. [CrossRef]
  107. (2024) — Playground v3: Improving Text-to-Image Alignment with Deep-Fusion Large Language Models https://arxiv.org/pdf/2409.10695. [CrossRef]
  108. (2023) — AI Art and its Impact on Artists https://dl.acm.org/doi/pdf/10.1145/3600211.3604681. [CrossRef]
  109. (2022) — AI Art in Architecture https://arxiv.org/pdf/2212.09399. [CrossRef]
  110. (2023) — AI Art and its Impact on Artists https://dl.acm.org/doi/pdf/10.1145/3600211.3604681. [CrossRef]
  111. (2023) — Google USM: Scaling Automatic Speech Recognition Beyond 100 Languages https://arxiv.org/pdf/2303.01037. [CrossRef]
  112. (2023) — Distil-Whisper: Robust Knowledge Distillation via Large-Scale Pseudo Labelling https://arxiv.org/pdf/2311.00430. [CrossRef]
  113. (2024) — Fine-tuning Whisper on Low-Resource Languages for Real-World Applications https://arxiv.org/pdf/2412.15726. [CrossRef]
  114. (2024) — OWSM v3.1: Better and Faster Open Whisper-Style Speech Models based on E-Branchformer https://arxiv.org/pdf/2401.16658. [CrossRef]
  115. (2024) — Transforming Conversations with AI—A Comprehensive Study of ChatGPT. [CrossRef]
  116. (2024) — Transforming Conversations with AI—A Comprehensive Study of ChatGPT. [CrossRef]
  117. (2024) — Systematic exploration and in-depth analysis of ChatGPT architectures progression. [CrossRef]
  118. (2024) — The Educational Affordances and Challenges of ChatGPT: State of the Field. [CrossRef]
  119. (2023) — ChatGPT and large language models in academia: opportunities and challenges. [CrossRef]
  120. (2024) — BLOOM: A 176B-Parameter Open-Access Multilingual Language Model https://arxiv.org/pdf/2211.05100. [CrossRef]
  121. (2024) — BLOOM: A 176B-Parameter Open-Access Multilingual Language Model https://arxiv.org/pdf/2211.05100. [CrossRef]
  122. (2024) — Towards a Framework for Openness in Foundation Models: Proceedings from the Columbia Convening on Openness in Artificial Intelligence https://arxiv.org/pdf/2405.15802. [CrossRef]
  123. (2024) — Near to Mid-term Risks and Opportunities of Open-Source Generative AI https://arxiv.org/pdf/2404.17047. [CrossRef]
  124. (2023) — Code Llama: Open Foundation Models for Code https://arxiv.org/pdf/2308.12950. [CrossRef]
  125. (2023) — LLaMA: Open and Efficient Foundation Language Models https://arxiv.org/pdf/2302.13971. [CrossRef]
  126. (2023) — Llama 2: Open Foundation and Fine-Tuned Chat Models https://arxiv.org/pdf/2307.09288. [CrossRef]
  127. (2024) — On the Societal Impact of Open Foundation Models https://arxiv.org/pdf/2403.07918. [CrossRef]
  128. (2024) — Google Gemini as a next generation AI educational tool: a review of emerging educational technology. [CrossRef]
  129. (2023) — Gemini: A Family of Highly Capable Multimodal Models https://arxiv.org/pdf/2312.11805. [CrossRef]
  130. (2024) — Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context https://arxiv.org/pdf/2403.05530. [CrossRef]
  131. (2025) — Comparative performance of neurosurgery-specific, peer-reviewed versus general AI chatbots in bilingual board examinations: evaluating accuracy, consistency, and error minimization strategies. [CrossRef]
  132. (2023) — Gemini: A Family of Highly Capable Multimodal Models https://arxiv.org/pdf/2312.11805. [CrossRef]
  133. (2023) — Frontier AI Regulation: Managing Emerging Risks to Public Safety https://arxiv.org/pdf/2307.03718. [CrossRef]
  134. (2024) — The Ethics of Advanced AI Assistants https://arxiv.org/pdf/2404.16244. [CrossRef]
  135. (2024) — Open Problems in Technical AI Governance https://arxiv.org/pdf/2407.14981. [CrossRef]
  136. (2025) — GPT-4o: The Cutting-Edge Advancement in Multimodal LLM. [CrossRef]
  137. (2024) — GPT-4o System Card https://arxiv.org/pdf/2410.21276. [CrossRef]
  138. (2025) — GPT-4o: The Cutting-Edge Advancement in Multimodal LLM. [CrossRef]
  139. (2024) — Mini-Omni2: Towards Open-source GPT-4o with Vision, Speech and Duplex Capabilities https://arxiv.org/pdf/2410.11190. [CrossRef]
  140. (2024) — From Efficient Multimodal Models to World Models: A Survey https://arxiv.org/pdf/2407.00118. [CrossRef]
  141. (2024) — Sora: A Review on Background, Technology, Limitations, and Opportunities of Large Vision Models https://arxiv.org/pdf/2402.17177. [CrossRef]
  142. (2024) — Sora: A Review on Background, Technology, Limitations, and Opportunities of Large Vision Models https://arxiv.org/pdf/2402.17177. [CrossRef]
  143. (2024) — Sora OpenAI’s Prelude: Social Media Perspectives on Sora OpenAI and the Future of AI Video Generation https://arxiv.org/pdf/2403.14665. [CrossRef]
  144. (2024) — Sora OpenAI’s Prelude: Social Media Perspectives on Sora OpenAI and the Future of AI Video Generation https://arxiv.org/pdf/2403.14665. [CrossRef]
  145. (2023) — Gemini: A Family of Highly Capable Multimodal Models https://arxiv.org/pdf/2312.11805. [CrossRef]
  146. (2023) — Gemini: A Family of Highly Capable Multimodal Models https://arxiv.org/pdf/2312.11805. [CrossRef]
  147. (2024) — Gemma: Open Models Based on Gemini Research and Technology https://arxiv.org/pdf/2403.08295. [CrossRef]
  148. (2024) — Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context https://arxiv.org/pdf/2403.05530. [CrossRef]
  149. (2024) — Mixtral of Experts https://arxiv.org/pdf/2401.04088. [CrossRef]
  150. (2023) — Mistral 7B https://arxiv.org/pdf/2310.06825. [CrossRef]
  151. (2024) — Linq-Embed-Mistral Technical Report https://arxiv.org/pdf/2412.03223. [CrossRef]
  152. (2023) — A Comprehensive Overview of Large Language Models https://arxiv.org/pdf/2307.06435. [CrossRef]
  153. (2023) — MemGPT: Towards LLMs as Operating Systems https://arxiv.org/pdf/2310.08560. [CrossRef]
  154. (2024) — The Dawn of GUI Agent: A Preliminary Case Study with Claude 3.5 Computer Use https://arxiv.org/pdf/2411.10323. [CrossRef]
  155. (2024) — Clio: Privacy-Preserving Insights into Real-World AI Use https://arxiv.org/pdf/2412.13678. [CrossRef]
  156. (2024) — DeepSeek-V3 Technical Report https://arxiv.org/pdf/2412.19437. [CrossRef]
  157. (2024) — DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model https://arxiv.org/pdf/2405.04434. [CrossRef]
  158. (2025) — DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models https://arxiv.org/pdf/2402.03300. [CrossRef]
  159. (2023) — VideoPoet: A Large Language Model for Zero-Shot Video Generation https://arxiv.org/pdf/2312.14125. [CrossRef]
  160. (2023) — VideoPoet: A Large Language Model for Zero-Shot Video Generation https://arxiv.org/pdf/2312.14125. [CrossRef]
  161. (2023) — Gemini: A Family of Highly Capable Multimodal Models https://arxiv.org/pdf/2312.11805. [CrossRef]
  162. (2024) — GPT-4o System Card https://arxiv.org/pdf/2410.21276. [CrossRef]
  163. (2024) — A Survey on Retrieval-Augmented Text Generation for Large Language Models https://arxiv.org/pdf/2404.10981. [CrossRef]
  164. (2025) — Singh, G. — *The Multiple Approaches for Drug-Drug Interaction Extraction using Machine Learning and Transformer-Based Model* bioRxiv, 2025-10. [CrossRef]
  165. (2025) — Singh, G. — *A Review of Multimodal Vision–Language Models: Foundations, Applications, and Future Directions* Preprints. [CrossRef]
  166. (2024) — Singh, G., Singh, S., Rehmani, N., Kumari, P., & Varshini, S. V. — *The Role of Data Analytics in Driving Business Innovation and Economic Growth: A Comparative Study Across Industries* International Journal of Innovative Research in Engineering and Management (IJIREM), Vol. 11, Issue 4, pp. 33–38. [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated