The Dartmouth Conference, formally known as the Dartmouth Summer Research Project on Artificial Intelligence, was a landmark academic workshop held in the summer of 1956 at Dartmouth College in Hanover, New Hampshire. Widely regarded as the founding event of artificial intelligence as a field, the workshop brought together a small group of researchers who believed that machines could be made to simulate human intelligence. It was at this conference that the term "artificial intelligence" was first formally used, giving a name to an entirely new discipline of computer science.
The workshop was organized by four scientists: John McCarthy, Marvin Minsky, Nathaniel Rochester, and Claude Shannon. Their 1955 proposal, submitted to the Rockefeller Foundation, laid out an ambitious research agenda that would shape the direction of AI research for decades to come.
By the mid-1950s, several threads of research had converged to suggest that building intelligent machines might be feasible. Claude Shannon had published foundational work on information theory in 1948. Alan Turing had introduced the concept of a universal computing machine and, in 1950, published his famous paper "Computing Machinery and Intelligence," which proposed what became known as the Turing test. Meanwhile, researchers at various institutions were beginning to write programs that could perform tasks previously thought to require human intelligence.
John McCarthy, then a young assistant professor of mathematics at Dartmouth, was the driving force behind the conference. He had become convinced that a focused summer workshop could accelerate progress in this new area. In 1955, McCarthy approached Marvin Minsky (then at Harvard), Nathaniel Rochester (a senior engineer at IBM who had designed the IBM 701), and Claude Shannon (at Bell Labs, already famous for information theory) to co-author a formal proposal.
The intellectual climate was ripe. Warren McCulloch and Walter Pitts had published their influential 1943 paper on a logical calculus of neural activity, laying theoretical groundwork for artificial neural networks. Donald Hebb's 1949 book The Organization of Behavior introduced what became known as Hebbian learning, a principle for how neural connections strengthen through use. And at the RAND Corporation, Allen Newell and Herbert Simon were already experimenting with computer programs that could reason symbolically. McCarthy recognized that these scattered efforts needed a unifying framework, a name, and a community [1].
The proposal, dated August 31, 1955, was submitted to the Rockefeller Foundation requesting funding for a two-month, ten-person study. The document was remarkably concise, running just over 600 words, yet it articulated a vision that would define an entire field. It opened with what has become one of the most quoted passages in AI history:
"The study is to proceed on the basis of the conjecture that every aspect of learning or any other feature of intelligence can in principle be so precisely described that a machine can be made to simulate it."
The proposal continued: "An attempt will be made to find how to make machines use language, form abstractions and concepts, solve kinds of problems now reserved for humans, and improve themselves. We think that a significant advance can be made in one or more of these problems if a carefully selected group of scientists work on it together for a summer" [1].
The proposal identified seven areas for investigation, each with its own brief justification. These sections reveal both the breadth of the organizers' ambitions and the specific intellectual commitments they brought to the project.
| Area | Description | Key Insight from Proposal |
|---|---|---|
| Automatic computers | Using computers to simulate aspects of intelligence | The proposal noted that computers were not yet fast enough for true intelligence simulation, but argued the main obstacle was programming, not hardware |
| Programming a computer to use language | Enabling machines to process and generate natural language | The authors suggested that "a large part of human thought consists of manipulating words according to rules of reasoning and rules of conjecture" |
| Neuron nets | Exploring how networks of neurons could be simulated to form concepts | The proposal acknowledged that "there is no theory yet to explain how it can be done," despite promising work by Farley and Clark on random nets |
| Theory of the size of a calculation | Understanding the computational complexity of problems | This section anticipated what would later become computational complexity theory, noting that some problems have a "certain intrinsic complexity" |
| Self-improvement | Building machines that could improve their own performance | The proposal envisioned machines that "could not only simulate any process of logical reasoning but also simultaneously improve the process" |
| Abstractions | Enabling machines to form abstractions and generalizations | The authors noted that programs "now in existence" could not form abstractions, calling this a key limitation |
| Randomness and creativity | Studying the relationship between randomness and creative thought | The proposal speculated on the "important and interesting connection between randomness and creative thinking," but also cautioned that randomness alone was insufficient |
Each of the four organizers contributed language to different sections of the proposal. McCarthy focused on language and abstraction, Shannon on information-theoretic aspects and the theory of calculation size, Rochester on neuron nets (drawing on his IBM experience with hardware simulation), and Minsky on self-improvement and learning [1][9].
The Rockefeller Foundation awarded a grant of $7,500 (equivalent to roughly $85,000 in 2024 dollars) to support the workshop. The organizers had originally requested $13,500 for the full program, so the grant covered just over half the proposed budget [1][10].
The four organizers each brought distinct expertise and perspectives to the project.
| Organizer | Affiliation | Age in 1956 | Contribution |
|---|---|---|---|
| John McCarthy | Dartmouth College | 29 | Coined the term "artificial intelligence"; later created Lisp programming language |
| Marvin Minsky | Harvard University (later MIT) | 29 | Pioneer of neural network theory and AI research |
| Nathaniel Rochester | IBM | 37 | Designed the IBM 701; worked on pattern recognition and geometry theorem proving |
| Claude Shannon | Bell Labs | 40 | Father of information theory; worked on chess-playing programs |
McCarthy deliberately chose the phrase "artificial intelligence" for the proposal. He later explained that he wanted a name that would distinguish this work from the broader field of cybernetics, which was associated with Norbert Wiener and had different intellectual commitments. Cybernetics tended to focus on feedback mechanisms and analog systems, while McCarthy envisioned a field centered on digital computation and symbolic reasoning. The term was somewhat controversial even at the time. Some participants, including Minsky, later expressed mixed feelings about the name, but it stuck and became the standard label for the field [2].
McCarthy also had strategic reasons for the naming. He later recalled that he "didn't want to see Norbert Wiener as the guru of the field," and that coining a new term helped establish intellectual independence from the cybernetics movement. This was not merely academic politics; it reflected genuine differences in approach. The cyberneticists emphasized continuous, analog processes inspired by biological feedback loops, while McCarthy and his colleagues were more interested in discrete, symbolic computation [6].
In addition to the four organizers, the workshop attracted several other researchers who would go on to make foundational contributions to computer science and AI.
| Attendee | Affiliation | Notable Work | Contribution at Dartmouth |
|---|---|---|---|
| Allen Newell | Carnegie Institute of Technology (now CMU) | Co-creator of the Logic Theorist and General Problem Solver | Presented the Logic Theorist, the first running AI program |
| Herbert Simon | Carnegie Institute of Technology | Co-creator of Logic Theorist; later won Nobel Prize in Economics (1978) | Co-presented Logic Theorist; argued for heuristic search as central to intelligence |
| Arthur Samuel | IBM | Created a checkers-playing program that learned from experience | Discussed his checkers program's ability to improve through self-play |
| Ray Solomonoff | Independent researcher | Founder of algorithmic probability and Solomonoff induction | Presented early ideas on machine learning and inductive inference; kept the most detailed notes of the workshop |
| Oliver Selfridge | MIT Lincoln Lab | Pioneer of machine perception; authored "Pandemonium" model | Discussed pattern recognition and his developing ideas about perception |
| Trenchard More | Princeton / MIT | Worked on formal language theory; co-explored the word "heuristic" with other attendees | Participated in discussions on language, heuristics, and formal reasoning |
Other visitors and participants also attended portions of the workshop, though exact attendance records are incomplete. The total number of people who spent time at the workshop was likely around 20, though the core group at any given moment was smaller. Ray Solomonoff's personal notes provide some of the most detailed accounts of daily activities and discussions at the workshop [3].
Each attendee brought a distinct research program that would blossom in subsequent decades:
Allen Newell and Herbert Simon arrived from Carnegie Tech with the most concrete achievement: a working program. Their collaboration with programmer J. Clifford Shaw at the RAND Corporation had produced the Logic Theorist, written in a new programming language they called IPL (Information Processing Language). Newell and Simon believed that intelligence was fundamentally about symbol manipulation and heuristic search, ideas that would dominate AI for the next 30 years.
Arthur Samuel had been working at IBM on a checkers-playing program since the early 1950s. His program was one of the earliest examples of what would later be called machine learning: it could improve its play through experience, using a form of rote learning combined with generalization. Samuel's approach was empirical and engineering-oriented, contrasting with the more theoretical inclinations of some other participants.
Ray Solomonoff was developing ideas about inductive inference and algorithmic probability that were well ahead of their time. His approach to machine learning, grounded in mathematical formalism, would not be fully appreciated for decades. At the workshop, he presented ideas about how machines might learn to predict sequences of data, laying groundwork for what would become algorithmic information theory [3].
Oliver Selfridge was working on pattern recognition at MIT's Lincoln Laboratory. His "Pandemonium" model, which he would publish in 1958, proposed a system of hierarchical "demons" that competed to interpret sensory input. This architecture anticipated key features of later neural network models.
The workshop ran for approximately six to eight weeks during the summer of 1956, from around June 18 to August 17. Contrary to the organizers' hopes for a concentrated collaboration among ten people working full-time, the reality was more informal and fragmented. Most participants came for only a week or two at a time, and the group was rarely all together at once. At its peak, there may have been only six or seven people present simultaneously [4].
The participants had the entire top floor of the Dartmouth Mathematics Department to themselves. Most weekdays, someone would lead a discussion focusing on their ideas, or more frequently, a general discussion would be held in the main math classroom. There was no fixed agenda or formal schedule. Instead, participants worked individually or in small groups, gathering for discussions and presentations as ideas arose. McCarthy later described the atmosphere as a "summer-long brainstorming session" rather than a structured conference [4][9].
The informality had both advantages and disadvantages. On the positive side, it allowed for deep, sustained conversations between researchers who might otherwise never have met. On the negative side, it meant that the workshop lacked the focused intensity the organizers had envisioned. As McCarthy later reflected, "anybody could come for whatever part of the time that he liked" [9].
Logic Theorist demonstration. Allen Newell and Herbert Simon, working with programmer Cliff Shaw at the RAND Corporation, arrived with what many historians consider the first true AI program: the Logic Theorist. This program was designed to prove theorems from Alfred North Whitehead and Bertrand Russell's Principia Mathematica. By the time of the workshop, it had successfully proved 38 of the first 52 theorems in Chapter 2 of that work. For one theorem (Theorem 2.85), it even found a proof that was more elegant than the one published by Russell and Whitehead. Newell and Simon were so proud of this result that they attempted to submit the proof to the Journal of Symbolic Logic with the Logic Theorist listed as a co-author, but the journal rejected the submission [5].
Despite its significance, the Logic Theorist received a surprisingly lukewarm reception at Dartmouth. Historian Pamela McCorduck later noted that "the evidence is that nobody save Newell and Simon themselves sensed the long-range significance of what they were doing." Simon himself later reflected that they were "probably fairly arrogant about it all" and observed the irony that "we already had done the first example of what they were after; and second, they didn't pay much attention to it" [6].
One reason for the tepid response may have been philosophical. Newell and Simon's approach was top-down and symbolic: they wrote explicit rules for logical reasoning. Other participants, particularly those interested in neural networks and learning, favored bottom-up approaches where intelligence would emerge from simpler mechanisms. This tension between symbolic and connectionist approaches would persist for decades in the AI community.
Discussions on search and heuristics. Several participants explored the idea of heuristic methods for problem solving. In a memorable anecdote, Trenchard More later recalled that "Selfridge, and Minsky, and McCarthy, and Ray Solomonoff, and I gathered around a dictionary on a stand to look up the word heuristic, because we thought that might be a useful word" [3]. The concept of heuristic search, finding good-enough solutions without exhaustively exploring every possibility, would become one of the central ideas in AI.
Early ideas about machine learning. Arthur Samuel discussed his checkers program, which could improve its play through experience. This work would later become one of the earliest and most influential examples of machine learning. Samuel's demonstration showed that a computer could get better at a task without being explicitly programmed with better strategies, a concept that seemed almost magical to some observers.
Neural network models. Discussions about how networks of simple elements could give rise to complex behavior were a recurring theme, reflecting the neuron-net topic in the original proposal. Rochester, drawing on his IBM hardware experience, discussed simulation of neural networks on digital computers.
Language and abstraction. McCarthy led discussions about how computers might be programmed to use natural language and form abstract concepts. These conversations helped shape his later development of the Lisp programming language, which he designed in 1958 specifically to support the kinds of symbolic reasoning discussed at Dartmouth.
Information-theoretic approaches. Shannon contributed perspectives from information theory, discussing how the mathematical framework of communication and coding might apply to problems of machine intelligence. His work on chess-playing programs also informed discussions about search strategies.
The Dartmouth Conference did not produce the dramatic breakthroughs its organizers had hoped for. There was no single publication, no unified theory, and no consensus on the best approach to building intelligent machines. In that narrow sense, the workshop fell short of its stated objectives.
However, its true significance was organizational and conceptual. The conference achieved several things that proved far more important than any specific technical result:
Established AI as a distinct field. Before Dartmouth, research on machine intelligence was scattered across cybernetics, automata theory, operations research, and various engineering disciplines. The conference gave this research a name, a community, and a sense of shared purpose.
Created a network of researchers. The personal connections formed at Dartmouth led to decades of collaboration and, sometimes, productive rivalry. McCarthy went on to found the AI lab at Stanford, while Minsky co-founded the AI lab at MIT. Newell and Simon built their program at Carnegie Mellon. These three institutions became the dominant centers of AI research for the next quarter century.
Defined the research agenda. The topics identified in the 1955 proposal, including natural language processing, learning, search, and knowledge representation, became the core subfields of AI research for the next several decades.
Demonstrated working AI programs. The Logic Theorist and Samuel's checkers program showed that AI was not merely theoretical speculation but a practical engineering endeavor.
Established competing paradigms. The workshop surfaced fundamental disagreements about the right approach to AI. The symbolic approach championed by Newell and Simon, the neural-network approach discussed by Rochester and others, and the mathematical-formalist approach represented by Solomonoff all had their advocates. These competing visions would drive the field's intellectual development for decades.
The Dartmouth workshop also set the tone for the ambitious, sometimes overconfident predictions that would characterize early AI research. The proposal itself suggested that "a significant advance" in one or more of the identified problems could be achieved during a single summer.
Herbert Simon famously predicted in 1957, shortly after the conference, that within ten years a computer would be chess champion of the world and that a computer would discover and prove an important new mathematical theorem. He and Newell also predicted in 1958 that within ten years a digital computer would compose music of aesthetic merit and that most theories in psychology would take the form of computer programs.
| Prediction | Who | When Made | Predicted Deadline | Actual Achievement |
|---|---|---|---|---|
| Computer chess champion | Herbert Simon | 1957 | By 1967 | Deep Blue defeated Kasparov in 1997 (40 years late) |
| Computer proves important math theorem | Herbert Simon | 1957 | By 1967 | Appel and Haken proved the four-color theorem with computer assistance in 1976 |
| Machine language translation | Dartmouth participants | 1956 | Within a few years | Still an active research area; neural machine translation achieved strong results in the 2010s |
| Machines compose quality music | Simon and Newell | 1958 | By 1968 | AI music composition remains an evolving field as of the 2020s |
| Machines understand spoken language | Dartmouth participants | 1956 | By 1970 | Speech recognition became practical in the 2010s with deep learning |
While chess would eventually fall to computers (with Deep Blue defeating Garry Kasparov in 1997), it took 41 years rather than ten.
This pattern of ambitious predictions followed by slower-than-expected progress became a recurring theme in AI history, contributing to periods of reduced funding and interest known as AI winters [7].
The Dartmouth Conference catalyzed a burst of activity in the years that immediately followed. The personal connections made at Dartmouth led directly to the establishment of the major AI research centers that would dominate the field's first generation.
Institutional formation. In 1958, McCarthy moved from Dartmouth to MIT, where he and Minsky co-founded the MIT Artificial Intelligence Project (which later became the MIT AI Lab, now CSAIL). In 1963, McCarthy left MIT for Stanford, where he founded the Stanford AI Lab (SAIL). At Carnegie Tech, Newell and Simon continued building on the Logic Theorist with the General Problem Solver (GPS) in 1959.
Key technical milestones of the first decade:
| Year | Milestone | Researchers | Connection to Dartmouth |
|---|---|---|---|
| 1956 | Logic Theorist | Newell, Simon, Shaw | Demonstrated at Dartmouth |
| 1958 | Lisp programming language | McCarthy | Inspired by Dartmouth discussions on language and abstraction |
| 1958 | Perceptron | Frank Rosenblatt | Built on neuron-net ideas discussed at Dartmouth |
| 1959 | General Problem Solver | Newell, Simon | Extended Logic Theorist's heuristic search approach |
| 1959 | Samuel's checkers program (improved) | Samuel | Refined the learning methods discussed at Dartmouth |
| 1961 | SAINT (Symbolic Automatic INTegrator) | James Slagle (Minsky student) | Applied symbolic AI to calculus problems |
| 1964 | ELIZA | Joseph Weizenbaum | Early natural language processing program at MIT |
| 1965 | Resolution principle | J. Alan Robinson | Advanced the theorem-proving agenda |
Funding and government support. The Dartmouth Conference helped attract government attention to AI. By the early 1960s, the Defense Advanced Research Projects Agency (DARPA) began funding AI research at MIT, Stanford, and Carnegie Mellon. This government support was crucial in sustaining research programs through the field's early years.
The symbolic AI paradigm. The approach that emerged dominant from the first decade was symbolic AI: programs that manipulated symbols according to explicit rules to solve problems. This paradigm, strongly associated with Newell and Simon's work, became the mainstream of AI research. The alternative connectionist approach (neural networks) experienced a period of decline after Minsky and Seymour Papert published Perceptrons in 1969, which highlighted the limitations of single-layer networks.
The Dartmouth Conference holds a unique place in the history of technology. It did not invent artificial intelligence in any technical sense; researchers had been working on related problems before the workshop, and the most impressive demonstration (the Logic Theorist) had been completed before the participants gathered. What Dartmouth accomplished was something arguably more important: it created a field.
By bringing together a critical mass of talented researchers, giving them a shared vocabulary and a common set of problems, the conference catalyzed a research program that would unfold over the following decades. The symbolic AI paradigm that dominated AI research from the 1950s through the 1980s has its intellectual roots in the discussions at Dartmouth.
The conference has been described as "the Constitutional Convention of AI" for its role in establishing the foundational principles and social structures of the field [8]. In 2006, Dartmouth College hosted a 50th anniversary celebration, the "AI@50" conference, which brought together surviving original participants with contemporary researchers to reflect on the field's progress. McCarthy, Minsky, Solomonoff, Selfridge, and More all attended the reunion. Sadly, Newell had passed away in 1992 and Shannon, though alive, was suffering from Alzheimer's disease and could not attend.
The Dartmouth Conference also represents an interesting case study in how scientific fields are born. Unlike many disciplines that emerge gradually through accumulated publications and shifting institutional structures, AI was, in a sense, declared into existence by a small group of researchers who decided that the problems they cared about deserved their own name and their own community. Whether this was an act of intellectual foresight or premature optimism remains a subject of debate among historians of science.
The conference illustrates the power of convening. The total investment was modest: $7,500 from the Rockefeller Foundation, a building to meet in, and six to eight weeks of time. Yet the returns on that investment, measured in terms of the intellectual and economic activity it catalyzed, have been incalculable. The AI industry that traces its lineage to that summer in Hanover generates hundreds of billions of dollars annually as of the 2020s.
Today, with AI technologies like deep learning, large language models, and reinforcement learning transforming industries worldwide, the vision articulated in that 1955 proposal continues to guide research. The conjecture that "every aspect of learning or any other feature of intelligence can in principle be so precisely described that a machine can be made to simulate it" remains both the field's foundational hypothesis and its greatest open question.