Artificial intelligence (AI) is the field of computer science dedicated to creating systems capable of performing tasks that normally require human intelligence. These tasks include learning from experience, recognizing patterns, understanding natural language, making decisions, and solving problems. The term also refers to the intelligence demonstrated by such systems, in contrast to the natural intelligence displayed by humans and other animals.
AI has grown from a niche academic pursuit in the 1950s into one of the most consequential technologies of the 21st century. As of 2026, AI systems power search engines, medical diagnostics, autonomous vehicles, scientific research tools, and conversational assistants used by hundreds of millions of people worldwide.
Imagine you have a really smart toy robot. You show it hundreds of pictures of cats and dogs, and every time you tell it which one is a cat and which one is a dog. After seeing enough pictures, the robot starts figuring out on its own which new pictures are cats and which are dogs, even pictures it has never seen before. That is basically what artificial intelligence does: it learns patterns from examples so it can make good guesses about new things. Some AI is really good at one specific job (like recognizing cats), and some researchers are trying to build AI that can learn lots of different jobs, just like a person can.
There is no single, universally accepted definition of artificial intelligence. Definitions have shifted over the decades, reflecting changing goals and capabilities within the field.
Stuart Russell and Peter Norvig, in their widely used textbook Artificial Intelligence: A Modern Approach, organize definitions along two dimensions: those concerned with thought processes versus behavior, and those measuring success against human performance versus ideal rationality. This yields four approaches: systems that think like humans, systems that act like humans, systems that think rationally, and systems that act rationally [1].
A practical working definition describes AI as the study and design of intelligent agents, where an intelligent agent is a system that perceives its environment and takes actions to maximize its chances of achieving its goals. Machine learning, the dominant subfield today, focuses on systems that improve their performance on a task through experience without being explicitly programmed for every scenario.
The boundary between "AI" and "ordinary software" is often debated. Tasks once considered hallmarks of AI, such as optical character recognition or chess playing, are sometimes excluded from the definition once they become routine. This phenomenon is informally called the "AI effect" [2].
The history of artificial intelligence spans more than seven decades, marked by periods of intense optimism, painful setbacks, and transformative breakthroughs.
The idea of artificial beings with human-like intelligence dates back to ancient myths and legends. In Greek mythology, Hephaestus crafted golden automatons to serve him. In the 17th century, Gottfried Wilhelm Leibniz and Thomas Hobbes explored the idea that rational thought could be reduced to mechanical calculation.
In 1936, Alan Turing published his landmark paper "On Computable Numbers," introducing the concept of the Turing machine, a theoretical device that formalized computation. This work laid the mathematical foundation for all of computer science, and by extension, for AI.
In 1950, Turing published "Computing Machinery and Intelligence" in the journal Mind, posing the question "Can machines think?" Rather than defining thought directly, he proposed the "imitation game" (now called the Turing test): a human evaluator engages in natural language conversation with both a human and a machine; if the evaluator cannot reliably distinguish the machine from the human, the machine is said to exhibit intelligent behavior [3]. This paper is widely regarded as the founding document of AI as a discipline.
The field of artificial intelligence was formally established at the Dartmouth Summer Research Project on Artificial Intelligence, held from June 18 to August 17, 1956, at Dartmouth College in Hanover, New Hampshire. The workshop was organized by John McCarthy, Marvin Minsky, Nathaniel Rochester, and Claude Shannon [4].
McCarthy coined the term "artificial intelligence" in the proposal for this conference, dated August 31, 1955. The proposal stated the foundational hypothesis: "every aspect of learning or any other feature of intelligence can in principle be so precisely described that a machine can be made to simulate it" [4]. The workshop attracted approximately ten participants, including Allen Newell and Herbert Simon, who presented the Logic Theorist, often considered the first AI program.
The years following Dartmouth saw rapid progress and high expectations. Key developments included:
During this period, researchers made optimistic predictions. Herbert Simon stated in 1965 that "machines will be capable, within twenty years, of doing any work a man can do." Minsky predicted in 1967 that "within a generation... the problem of creating 'artificial intelligence' will substantially be solved" [2]. These predictions proved wildly premature.
By the early 1970s, it became clear that AI had not delivered on its grand promises. Fundamental problems proved far harder than expected, including natural language understanding, common-sense reasoning, and the "combinatorial explosion" of possibilities that made many problems computationally intractable.
In 1973, the British mathematician Sir James Lighthill published a report for the UK Science Research Council that criticized AI research for failing to achieve its "grandiose objectives." The Lighthill Report led to the near-complete dismantling of AI research in the United Kingdom [6]. In the United States, DARPA significantly reduced its funding for AI research starting in 1974. This period of reduced funding and diminished interest became known as the first "AI winter."
AI experienced a commercial resurgence in the early 1980s through expert systems, programs that emulated the decision-making ability of human specialists by encoding domain knowledge as "if-then" rules.
R1 (also known as XCON), developed at Carnegie Mellon University for Digital Equipment Corporation starting in 1978, became one of the first successful commercial expert systems. By 1986, it was saving DEC an estimated $40 million per year. Japan's Fifth Generation Computer Systems project, launched in 1982, aimed to build massively parallel computers optimized for AI, spurring the United States and United Kingdom to increase their own AI funding in response [2].
The expert systems market grew to over $1 billion by the mid-1980s. Companies like Symbolics, Lisp Machines Inc., and Texas Instruments sold specialized AI hardware.
The expert systems bubble burst beginning around 1987. Expert systems proved expensive to maintain, difficult to update, and brittle when faced with situations outside their narrow rule sets. The specialized Lisp machines became obsolete as conventional desktop computers grew more powerful. The Japanese Fifth Generation project failed to meet its ambitious goals. By 1993, over 300 AI companies had shut down, been acquired, or quietly pivoted away from AI [2].
Funding dried up again, and "artificial intelligence" became something of a stigmatized term in the tech industry. Many researchers rebranded their work under labels like "informatics," "knowledge systems," or "computational intelligence."
Despite the public chill, important advances continued.
On May 11, 1997, IBM's Deep Blue defeated reigning world chess champion Garry Kasparov 3.5 to 2.5 in a six-game match in New York City. It was the first time a computer had beaten a world champion under standard tournament conditions. Deep Blue could evaluate approximately 200 million positions per second [7]. While Deep Blue relied on brute-force search rather than general intelligence, the event captured worldwide attention and demonstrated the growing power of specialized AI.
In February 2011, IBM's Watson defeated the two all-time greatest Jeopardy! champions, Ken Jennings and Brad Rutter, in a televised match viewed by millions. Watson was a room-sized system consisting of 90 servers with 2,880 processor cores, and it could understand questions posed in natural language and retrieve answers without an internet connection. The victory demonstrated significant advances in natural language processing and open-domain question answering [8].
During the late 1990s and 2000s, the field shifted toward statistical methods and data-driven approaches. Advances in machine learning, including support vector machines, random forests, and Bayesian methods, showed that many tasks could be handled more effectively by learning patterns from large datasets than by encoding rules manually. The growing availability of digital data and increasing computational power fueled this transition.
The modern era of AI began with the deep learning breakthrough. In September 2012, Alex Krizhevsky, Ilya Sutskever, and Geoffrey Hinton demonstrated AlexNet, a deep convolutional neural network that achieved a top-5 error rate of 15.3% on the ImageNet image recognition challenge, compared to the previous best of 26.2%. This dramatic improvement proved that deep neural networks, trained on GPUs with large datasets, could vastly outperform hand-engineered approaches in computer vision [9].
AlexNet triggered a wave of investment and research in deep learning. Key milestones that followed include:
| Year | Milestone | Significance |
|---|---|---|
| 2014 | Generative adversarial networks (GANs) introduced by Ian Goodfellow | Enabled realistic image generation and data augmentation |
| 2014 | DeepFace by Facebook achieves near-human face recognition | 97.35% accuracy on Labeled Faces in the Wild benchmark |
| 2015 | ResNet wins ImageNet with 152-layer deep network | Demonstrated that much deeper networks could be trained effectively |
| 2016 | AlphaGo defeats Lee Sedol 4-1 in Go | First program to beat a top professional Go player, watched by 200 million people [10] |
| 2017 | "Attention Is All You Need" paper introduces the transformer architecture | Became the foundation for virtually all modern large language models [11] |
| 2018 | BERT (Bidirectional Encoder Representations from Transformers) released by Google | Set new benchmarks across 11 NLP tasks simultaneously |
| 2018 | OpenAI releases GPT-1 | Demonstrated unsupervised pre-training for language understanding |
The victory of Google DeepMind's AlphaGo over Lee Sedol in March 2016 was particularly significant. Go had long been considered too complex for computers due to its vast search space (approximately 10^170 possible board positions). AlphaGo combined deep neural networks with Monte Carlo tree search to overcome this challenge [10].
The transformer architecture, proposed by Ashish Vaswani and colleagues at Google in June 2017, replaced the recurrent neural networks that had previously dominated sequence modeling. Its key innovation, the self-attention mechanism, allowed the model to weigh the relevance of all parts of an input simultaneously, enabling massive parallelization during training. The transformer became the architectural basis for GPT, BERT, and nearly every major language model that followed [11].
In June 2020, OpenAI released GPT-3, a large language model with 175 billion parameters trained on a vast corpus of internet text. GPT-3 demonstrated an unprecedented ability to generate coherent text, translate languages, answer questions, and even write code, all from natural language prompts. It was made available through an API, enabling thousands of applications [12].
In November 2020, Google DeepMind's AlphaFold 2 solved the protein structure prediction problem at the 14th CASP competition with unprecedented accuracy, predicting the three-dimensional structures of proteins from amino acid sequences alone. This breakthrough was later recognized with the 2024 Nobel Prize in Chemistry, awarded to Demis Hassabis, John Jumper, and David Baker [13].
On November 30, 2022, OpenAI launched ChatGPT, a conversational interface built on GPT-3.5 that was fine-tuned using reinforcement learning from human feedback (RLHF). ChatGPT reached 100 million monthly active users within two months, making it the fastest-growing consumer application in history at that time [14]. Its launch triggered a global wave of interest, investment, and concern about AI.
In March 2023, OpenAI released GPT-4, a multimodal model capable of processing both text and images, which demonstrated substantial improvements in reasoning, factual accuracy, and performance on professional examinations.
In October 2024, the Nobel Prizes recognized AI in two categories. John J. Hopfield and Geoffrey Hinton received the Nobel Prize in Physics "for foundational discoveries and inventions that enable machine learning with artificial neural networks." Demis Hassabis, John Jumper, and David Baker received the Nobel Prize in Chemistry for computational protein design and structure prediction using AI [15]. These awards underscored the transformative scientific impact of artificial intelligence.
The period from 2023 onward saw an explosion of generative AI development across the industry, including image generators like DALL-E, Midjourney, and Stable Diffusion; code assistants like GitHub Copilot; and competing large language models from Google (Gemini), Anthropic (Claude), Meta (LLaMA), and others.
AI systems are commonly categorized by the breadth of their capabilities.
Narrow AI, also called weak AI, refers to systems designed and trained to perform a specific task or a limited range of tasks. Every AI system in operation today falls into this category. Examples include spam filters, recommendation algorithms, image classifiers, voice assistants, and large language models.
Narrow AI can equal or exceed human performance within its designated domain. A chess engine can beat any human player, and modern language models can pass bar examinations. But these systems cannot transfer their abilities to unrelated tasks. A chess engine cannot drive a car, and a language model cannot physically manipulate objects.
Artificial general intelligence (AGI) refers to a hypothetical AI system with the ability to understand, learn, and apply knowledge across the full range of cognitive tasks that a human can perform. An AGI system would be able to reason abstractly, transfer knowledge between domains, learn from limited examples, and handle novel situations without specific training.
As of 2026, AGI does not exist. There is significant debate about when, or whether, it will be achieved. Some researchers and industry leaders, including Sam Altman of OpenAI, have suggested that AGI could be achieved within years. Others, including many academic AI researchers, argue that current approaches based on scaling large language models are insufficient and that fundamental new insights will be needed [16].
Superintelligence refers to a hypothetical AI that surpasses the cognitive abilities of the brightest humans in virtually every domain, including scientific reasoning, social intelligence, and creativity. The concept was extensively analyzed by philosopher Nick Bostrom in his 2014 book Superintelligence: Paths, Dangers, Strategies.
Superintelligence remains firmly in the realm of speculation. Its significance lies primarily in its role in debates about AI safety and existential risk, where researchers consider how to ensure that extremely capable AI systems remain aligned with human values and intentions.
Artificial intelligence encompasses a broad array of research areas and technical disciplines.
| Subfield | Description | Key applications |
|---|---|---|
| Machine learning | Systems that learn from data to make predictions or decisions without explicit programming | Recommendation systems, fraud detection, medical diagnosis |
| Deep learning | Subset of ML using multi-layered neural networks to learn hierarchical representations | Image recognition, speech synthesis, language modeling |
| Natural language processing (NLP) | Processing, understanding, and generating human language | Chatbots, translation, text summarization, sentiment analysis |
| Computer vision | Enabling machines to interpret and understand visual information from images and video | Autonomous driving, medical imaging, facial recognition |
| Robotics | Design and operation of robots, often integrating AI for perception, planning, and control | Manufacturing, surgery, warehouse automation, exploration |
| Expert systems | Rule-based programs that emulate human specialist decision-making | Medical diagnosis, financial planning, industrial process control |
| Knowledge representation | Methods for encoding information about the world in forms usable by AI systems | Ontologies, semantic web, knowledge graphs |
| Planning and scheduling | Algorithms for generating sequences of actions to achieve goals | Logistics, resource allocation, game playing |
| Speech recognition | Converting spoken language into text | Voice assistants, dictation software, accessibility tools |
Over the decades, researchers have pursued fundamentally different strategies for building intelligent systems.
Symbolic AI, sometimes called "Good Old-Fashioned AI" (GOFAI), was the dominant paradigm from the 1950s through the 1980s. It represents knowledge using human-readable symbols and manipulates these symbols according to explicit rules. Logic programming, expert systems, and semantic networks are all examples of symbolic approaches.
Symbolic AI excels at tasks involving structured reasoning, formal logic, and domains where knowledge can be clearly articulated. Its weaknesses include difficulty handling uncertainty, poor performance on perception tasks (like vision and speech), and the labor-intensive process of manually encoding knowledge (the "knowledge acquisition bottleneck") [17].
Connectionism models intelligence using artificial neural networks inspired (loosely) by the structure of biological brains. Rather than encoding explicit rules, connectionist systems learn patterns from data by adjusting the strengths of connections between large numbers of simple processing units (neurons).
The approach dates to Warren McCulloch and Walter Pitts's 1943 model of artificial neurons. It fell out of favor after the publication of Perceptrons in 1969, experienced a resurgence in the 1980s when David Rumelhart, Geoffrey Hinton, and Ronald Williams popularized the backpropagation algorithm, and has dominated AI since the deep learning breakthrough of 2012 [5].
Statistical AI uses probability theory and statistical methods to handle uncertainty and learn from data. Bayesian networks, hidden Markov models, and support vector machines fall into this category. These approaches became increasingly prominent in the 1990s and 2000s as alternatives to both symbolic and neural methods.
Modern AI systems increasingly combine multiple approaches. Neurosymbolic AI, for instance, integrates neural networks (for perception and pattern recognition) with symbolic reasoning (for logical inference and explainability). AlphaGo itself was a hybrid system, combining deep neural networks with Monte Carlo tree search, a symbolic planning technique [10].
Supervised learning trains a model on labeled data, where each input is paired with the correct output. The model learns to map inputs to outputs and can then make predictions on new, unseen data. Classification (assigning labels) and regression (predicting continuous values) are the two main types. Common algorithms include linear regression, decision trees, random forests, and neural networks.
Unsupervised learning works with unlabeled data, seeking to find hidden structure or patterns. Clustering (grouping similar data points) and dimensionality reduction (compressing data while preserving key information) are typical tasks. K-means clustering, principal component analysis (PCA), and autoencoders are standard techniques.
Reinforcement learning (RL) trains an agent to make sequences of decisions by rewarding desired behaviors and penalizing undesired ones. The agent learns a policy that maximizes cumulative reward over time through trial and error. RL has achieved notable successes in game playing (Atari, Go, StarCraft II), robotics, and fine-tuning language models through RLHF.
Artificial neural networks consist of layers of interconnected nodes (neurons) that process information. A network with more than two hidden layers is generally called a "deep" neural network, and training such networks is called deep learning.
Key architectures include:
Transfer learning involves taking a model trained on one task and adapting it for a different but related task. Foundation models are large models (often with billions of parameters) pre-trained on broad data that can be adapted to many downstream tasks through fine-tuning or prompting. GPT-4, Claude, Gemini, and LLaMA are all foundation models. This paradigm has transformed AI by making state-of-the-art capabilities accessible without requiring training from scratch [18].
The following table summarizes key milestones in the history of artificial intelligence.
| Year | Milestone | Significance |
|---|---|---|
| 1950 | Turing publishes "Computing Machinery and Intelligence" | Proposed the imitation game (Turing test) as a measure of machine intelligence |
| 1956 | Dartmouth Conference | Formal founding of AI as an academic field; the term "artificial intelligence" coined |
| 1958 | Frank Rosenblatt builds the Perceptron | First hardware implementation of an artificial neural network |
| 1966 | ELIZA created by Weizenbaum | Early demonstration of natural language interaction via pattern matching |
| 1997 | Deep Blue defeats Kasparov in chess | First computer to beat a reigning world chess champion under standard conditions |
| 2011 | Watson wins Jeopardy! | Demonstrated open-domain question answering in natural language |
| 2012 | AlexNet wins ImageNet challenge | Sparked the deep learning revolution in computer vision |
| 2016 | AlphaGo defeats Lee Sedol in Go | Conquered the last major classical board game considered too complex for computers |
| 2017 | Transformer architecture introduced | Foundation of all major modern large language models |
| 2020 | GPT-3 released; AlphaFold 2 solves protein folding | Milestone in language generation; breakthrough in computational biology |
| 2022 | ChatGPT launched | Reached 100 million users in two months; catalyzed global generative AI adoption |
| 2023 | GPT-4 released | Multimodal capabilities with strong reasoning across text and images |
| 2024 | AI recognized with two Nobel Prizes | Physics (Hopfield, Hinton) and Chemistry (Hassabis, Jumper, Baker) |
AI has found applications across virtually every major industry.
AI tools analyze medical images (X-rays, MRIs, CT scans, pathology slides) with accuracy that sometimes matches or exceeds specialist physicians. Between 2020 and 2023, the AI healthcare market expanded by 233%, with 94% of healthcare companies reporting some use of AI or machine learning [19].
Specific applications include drug discovery (predicting molecular interactions to identify candidate therapies), clinical decision support, administrative automation (reducing documentation time), and wearable health monitors that use AI to detect anomalies in vital signs. Google DeepMind's AlphaFold, released in 2020, predicted the three-dimensional structures of nearly all known proteins, a breakthrough recognized with the 2024 Nobel Prize in Chemistry [13].
Financial institutions use AI for algorithmic trading, credit scoring, fraud detection (analyzing transaction patterns in real time), risk assessment, customer service chatbots, and regulatory compliance. The AI-in-finance market was projected to generate approximately $450 billion by 2025 [19].
Self-driving cars and trucks rely on AI for perception (processing data from cameras, lidar, and radar), planning (determining routes and maneuvers), and control (executing driving actions). Waymo, a subsidiary of Alphabet, reported providing over 150,000 autonomous rides per week in the United States as of 2025. Baidu's Apollo Go robotaxi service has expanded across multiple cities in China [19].
AI accelerates research in physics, chemistry, biology, materials science, and mathematics. Beyond protein folding, AI systems have been used to discover new materials, predict weather patterns with greater accuracy than traditional models, find new mathematical proofs, and analyze vast astronomical datasets.
AI-powered code assistants like GitHub Copilot, Cursor, and Claude Code help programmers write, debug, review, and refactor code. These tools use large language models trained on open-source code repositories. Studies have reported productivity gains of 20% to 55% for certain programming tasks [20].
Generative AI creates text, images, audio, music, and video from natural language descriptions. Text generation is handled by models like GPT-4, Claude, and Gemini. Image generation uses models such as DALL-E, Midjourney, and Stable Diffusion. Video generation tools include Sora (OpenAI) and Veo (Google). These capabilities have broad applications in marketing, entertainment, education, and design, while also raising concerns about misinformation and intellectual property.
AI is increasingly used in education for personalized tutoring, adaptive learning platforms, automated grading, and language learning applications. Intelligent tutoring systems can adjust the difficulty of material in real time based on student performance. AI-powered tools assist educators with curriculum design, content generation, and identifying students who may need additional support.
The rapid advancement of AI has intensified longstanding ethical debates and introduced new ones.
AI systems can inherit and amplify biases present in their training data. Documented examples include facial recognition systems that perform poorly on darker-skinned individuals, hiring algorithms that discriminate against women, and language models that reproduce racial stereotypes. Addressing algorithmic bias requires diverse training data, careful evaluation across demographic groups, and ongoing monitoring after deployment [21].
AI enables large-scale surveillance, facial recognition in public spaces, and inference of personal attributes from digital footprints. These capabilities raise significant concerns about individual privacy and civil liberties. The use of AI in predictive policing, social credit systems, and targeted advertising has drawn particular criticism.
The World Economic Forum's Future of Jobs Report 2025 projected that 92 million jobs will be displaced by AI and automation by 2030, while 170 million new jobs will be created, yielding a net gain of 78 million jobs. However, the transition is expected to disproportionately affect low-skill and repetitive roles, with entry-level administrative positions seeing declines of around 35%. Without intervention through retraining and education, automation may deepen economic inequality [22].
The environmental cost of training and running large AI models has become a growing concern. The carbon footprint of AI systems was estimated at between 32.6 and 79.7 million tons of CO2 emissions in 2025, while the water footprint could reach 312 to 765 billion liters. The International Energy Agency estimated that AI systems accounted for 15% of total data center electricity demand in 2024 and projected that overall data center energy demand will double by 2030 due in large part to AI growth. An August 2025 analysis from Goldman Sachs Research forecast that approximately 60% of increasing electricity demands from data centers would be met by burning fossil fuels, adding roughly 220 million tons of carbon emissions globally [23].
Generative AI can produce highly realistic synthetic text, images, audio, and video, commonly called deepfakes. These capabilities create risks for misinformation, political manipulation, fraud, and erosion of public trust. Researchers and policymakers have proposed technical countermeasures such as digital watermarking, content provenance standards (like C2PA), and detection tools, though the effectiveness of these measures remains an open challenge.
AI safety is the research field dedicated to ensuring that AI systems operate reliably and do not cause unintended harm. AI alignment, a core subproblem, focuses on ensuring that an AI system's goals and behaviors remain consistent with human values and intentions as its capabilities increase.
The alignment problem arises because specifying human values and preferences precisely enough for an AI system to follow them faithfully is extremely difficult. A system that optimizes for a poorly specified objective can produce outcomes that satisfy the literal goal while violating the intended spirit. This challenge becomes more acute as AI systems become more capable, since a more powerful optimizer can find more unexpected (and potentially harmful) ways to achieve a misspecified goal.
Key areas of active safety research include:
Some researchers and public figures have warned that sufficiently advanced AI could pose an existential threat to humanity. In May 2023, the Center for AI Safety released a one-sentence statement, signed by hundreds of AI researchers and public figures including Geoffrey Hinton and Yoshua Bengio, asserting that "mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war" [25]. Not all experts agree on the severity of this risk. Critics argue that focusing on speculative long-term threats distracts from more immediate harms such as bias, misinformation, and environmental impact.
The 2026 International AI Safety Report warned that reliable safety testing has become increasingly difficult as models learn to distinguish between test environments and real-world deployment, raising concerns that pre-deployment evaluations may not accurately predict real-world behavior [24].
Governments around the world have begun developing regulatory frameworks for artificial intelligence, balancing innovation with risk mitigation.
The European Union's AI Act, adopted on May 21, 2024, is the first comprehensive AI regulation by a major jurisdiction. It employs a risk-based framework with four tiers [26]:
| Risk level | Description | Examples | Requirements |
|---|---|---|---|
| Unacceptable | AI practices deemed too harmful to permit | Social scoring by governments, real-time biometric surveillance in public spaces (with limited exceptions) | Banned entirely |
| High | AI systems that significantly affect safety or fundamental rights | Medical devices, hiring tools, credit scoring, law enforcement tools | Conformity assessments, transparency, human oversight |
| Limited | AI systems with specific transparency risks | Chatbots, deepfakes | Users must be informed they are interacting with AI |
| Minimal | Low-risk AI applications | Spam filters, AI-enabled video games | No specific requirements |
Key implementation dates include February 2, 2025 (prohibited practices and AI literacy requirements took effect), August 2, 2025 (governance rules and obligations for general-purpose AI models became applicable), and August 2, 2026 (high-risk AI system rules take effect) [26].
On October 30, 2023, President Biden signed Executive Order 14110, titled "Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence." It required companies developing high-performance AI models to notify the government and report results of red-team safety tests, streamlined visa processes for AI talent, established a pilot of the National AI Research Resource, and directed federal agencies to appoint Chief Artificial Intelligence Officers [27].
On January 20, 2025, President Trump revoked Executive Order 14110 on his first day in office. Three days later, he signed Executive Order 14179, "Removing Barriers to American Leadership in Artificial Intelligence," which shifted the federal approach from oversight and risk mitigation toward deregulation and promotion of AI innovation [28].
The National Institute of Standards and Technology (NIST) released the AI Risk Management Framework (AI RMF 1.0) in January 2023. The framework is voluntary, sector-agnostic, and organized around four core functions: Govern, Map, Measure, and Manage. In July 2024, NIST released a companion Generative AI Profile (NIST-AI-600-1) with specific guidance for managing risks from generative AI systems. The NIST AI RMF has become one of the most widely referenced voluntary governance frameworks for AI globally [29].
China's Interim Measures for the Management of Generative AI Services took effect in August 2023, requiring providers to ensure the truthfulness, accuracy, and lawfulness of training data. The United Kingdom has pursued a "pro-innovation" approach, distributing regulatory responsibility across existing sector-specific regulators rather than creating new AI-specific legislation.
Artificial intelligence raises deep philosophical questions about the nature of mind, understanding, and consciousness.
In 1980, philosopher John Searle published "Minds, Brains, and Programs," introducing the Chinese Room thought experiment. Searle imagines himself sitting in a room, following a manual to match incoming Chinese characters with appropriate Chinese responses. To an outside observer, the room appears to understand Chinese, but Searle (inside the room) does not understand a word of it. He merely manipulates symbols according to rules.
Searle argued that this demonstrates that a computer running a program cannot have genuine understanding or intentionality, no matter how convincingly it produces correct outputs. The argument targets "strong AI," the claim that an appropriately programmed computer literally possesses a mind. The Chinese Room remains one of the most debated thought experiments in philosophy of mind, with numerous responses including the "systems reply" (the whole room, not just the person inside, understands Chinese) and the "robot reply" (a system grounded in the physical world through sensors might achieve understanding) [30].
Whether AI systems can be conscious is an open and contentious question. A 2023 framework published in Trends in Cognitive Sciences by researchers including Yoshua Bengio and philosopher David Chalmers assessed AI systems against indicators derived from leading neuroscientific theories of consciousness. The authors concluded that "no current AI systems are conscious" but noted that "there are no obvious technical barriers to building AI systems which satisfy these indicators" [31].
As of 2025, large language models can pass behavioral tests of intelligence such as the Turing test (OpenAI's GPT-4.5 was judged as human 73% of the time in one study), but passing behavioral tests does not settle the question of conscious experience. The distinction between simulating intelligence and actually possessing subjective experience remains a central challenge for both philosophy and AI research.
The frame problem, originally formulated in the context of symbolic AI, concerns how an AI system can efficiently determine which aspects of its knowledge remain unchanged when an action is performed. More broadly, it raises the question of how an intelligent agent can determine what is relevant in a given situation. Humans effortlessly filter irrelevant information, but formalizing this ability for AI systems has proven remarkably difficult. The frame problem has motivated significant work in knowledge representation, common-sense reasoning, and cognitive architecture design.
The AI field is advancing at an extraordinary pace, driven by massive investment, fierce competition among technology companies, and rapidly improving model capabilities.
The leading AI models as of early 2026 include:
| Model | Developer | Release | Notable capabilities |
|---|---|---|---|
| GPT-5 | OpenAI | August 2025 | Multimodal reasoning, significantly reduced hallucination rates (up to 80% fewer factual errors than GPT-4) |
| Claude Sonnet 5 | Anthropic | February 2026 | 82.1% on SWE-Bench Verified (first model to break 80%), strong multi-step reasoning |
| Claude Opus 4.6 / Sonnet 4.6 | Anthropic | Early 2026 | 1-million-token context window |
| Gemini 2.5 | Google DeepMind | Late 2025 | Strong reasoning capabilities, LMArena benchmark leader |
| LLaMA 4 | Meta AI | 2025 | Open-weight model with strong tool orchestration |
| Grok | xAI | 2025 | Integrated with X (formerly Twitter) platform |
These models are increasingly multimodal, processing and generating text, images, audio, and video within single systems.
Agentic AI emerged as a major trend in 2025 and 2026. Unlike traditional AI assistants that respond to individual prompts, agentic systems can autonomously perceive their environment, formulate multi-step plans, execute actions, use external tools, and adapt based on outcomes, all with minimal human intervention between steps [32].
McKinsey's State of AI 2025 report found that 62% of organizations were actively working with AI agents. Gartner projected that 15% of day-to-day work decisions would be made autonomously through agentic AI by 2028, up from less than 1% in 2024 [32]. Anthropic's Model Context Protocol (MCP), released in late 2024, became a widely adopted open standard for connecting AI agents to external tools and data sources, with OpenAI and Microsoft publicly embracing the protocol in 2025.
AI infrastructure spending has reached unprecedented levels. The major hyperscale cloud providers collectively planned to spend nearly $700 billion on data center projects in 2026, roughly double the approximately $365 billion spent in 2025. Individual company capital expenditure plans for 2026 include Amazon at approximately $200 billion, Google at $175 billion to $185 billion, and Meta at $115 billion to $135 billion [33].
NVIDIA, the dominant supplier of AI training and inference chips, announced a partnership with OpenAI to deploy at least 10 gigawatts of NVIDIA systems. NVIDIA CEO Jensen Huang stated in March 2026 that he saw "at least $1 trillion" in computing demand through 2027 [33].
A number of companies and laboratories lead global AI research and development.
| Organization | Founded | Headquarters | Focus and notable contributions |
|---|---|---|---|
| OpenAI | 2015 | San Francisco, USA | GPT series, ChatGPT, DALL-E, Sora; surpassed $25 billion in annualized revenue by early 2026 |
| Google DeepMind | 2010 (DeepMind); merged with Google Brain in 2023 | London, UK | AlphaGo, AlphaFold, Gemini; two Nobel Prize-winning breakthroughs |
| Anthropic | 2021 | San Francisco, USA | Claude models, Constitutional AI, Model Context Protocol (MCP); focused on AI safety research |
| Meta AI (FAIR) | 2013 | Menlo Park, USA | LLaMA open-weight models, PyTorch framework, foundational research in self-supervised learning |
| Microsoft Research | 1991 | Redmond, USA | Copilot, Azure AI; major investor in OpenAI; integrated AI across Office suite |
| NVIDIA | 1993 | Santa Clara, USA | H100, B200, Blackwell GPUs, CUDA; dominant hardware supplier for AI training and inference |
| xAI | 2023 | San Francisco, USA | Grok models; founded by Elon Musk |
| Mistral AI | 2023 | Paris, France | Mistral Large, open-weight models; leading European AI company |
| Allen Institute for AI (AI2) | 2014 | Seattle, USA | OLMo open language models, Semantic Scholar; nonprofit research focus |
| EleutherAI | 2020 | Distributed | GPT-NeoX, The Pile dataset; open-source, grassroots AI research collective |
Academic institutions also play a critical role, with Stanford HAI, MIT CSAIL, UC Berkeley's BAIR, Carnegie Mellon's School of Computer Science, the University of Toronto, and Mila (Quebec AI Institute, co-founded by Yoshua Bengio) among the most influential.