Andrej Karpathy (born October 23, 1986) is a Slovak-Canadian computer scientist, AI researcher, educator, and entrepreneur. He is one of the most recognized figures in the artificial intelligence community, known for his ability to explain complex deep learning concepts in accessible terms and for his influential roles at OpenAI and Tesla. Karpathy was a founding member of OpenAI in 2015, served as Tesla's Senior Director of AI and head of Autopilot Vision from 2017 to 2022, briefly returned to OpenAI in 2023, and has since worked independently as an educator, content creator, and founder of Eureka Labs, an AI-native education startup [1].
Karpathy's academic contributions include co-creating the Stanford course CS231n: Convolutional Neural Networks for Visual Recognition, which became one of the most popular computer science courses at the university and introduced thousands of students to deep learning. His open-source implementations of language models, particularly minGPT, nanoGPT, and llm.c, have been widely used as educational tools. In February 2025, he coined the term "vibe coding" to describe a new style of programming in which developers rely on AI assistants to generate code from natural language descriptions rather than writing it manually. The term entered mainstream vocabulary rapidly and was named the Collins English Dictionary Word of the Year for 2025 [2]. In 2024, TIME magazine named Karpathy to its list of the 100 Most Influential People in AI [3].
Andrej Karpathy was born on October 23, 1986, in Bratislava, Czechoslovakia (now Slovakia). His family moved to Toronto, Canada, when he was 15 years old. He completed his secondary education in Toronto and went on to study at the University of Toronto, where he earned a bachelor's degree in Computer Science and Physics, with a minor in Mathematics, in 2009. While an undergraduate at the University of Toronto, Karpathy attended Geoffrey Hinton's class and participated in Hinton's reading groups, an experience that first exposed him to neural networks and deep learning and proved formative for his career [1][4].
Karpathy continued his graduate studies at the University of British Columbia (UBC), where he received a master's degree in Computer Science in 2011. His master's research, supervised by Michiel van de Panne, focused on curriculum learning for physically simulated characters. The work explored how simulated agents could acquire complex motor skills through staged, incremental learning, drawing inspiration from how humans and animals develop physical abilities in nature. His thesis, titled "Staged Learning of Agile Motor Skills," applied these techniques to planar characters learning skills such as hopping, flipping, rolling, and continuous acrobatic movements [5].
Karpathy then moved to Stanford University for his PhD, which he completed in 2015 under the supervision of Fei-Fei Li, one of the leading researchers in computer vision and the creator of the ImageNet dataset. His doctoral thesis, titled "Connecting Images and Natural Language," focused on the intersection of computer vision and natural language processing, developing deep learning models that could generate natural language descriptions of images and video content. The thesis synthesized several lines of research into scalable neural network architectures for processing visual-linguistic data, including image captioning, dense captioning, and video understanding [6].
| Milestone | Year | Details |
|---|---|---|
| Born | 1986 | Bratislava, Czechoslovakia (now Slovakia) |
| Moved to Canada | ~2001 | Settled in Toronto |
| BSc Computer Science and Physics (minor in Mathematics), University of Toronto | 2009 | Attended Geoffrey Hinton's class and reading groups |
| MSc Computer Science, University of British Columbia | 2011 | Advisor: Michiel van de Panne; thesis on physically simulated characters |
| PhD Computer Science, Stanford University | 2015 | Advisor: Fei-Fei Li; thesis: "Connecting Images and Natural Language" |
During his PhD at Stanford, Karpathy designed and became the lead instructor of CS231n, a course on convolutional neural networks for visual recognition. The course was one of the first dedicated deep learning courses offered at a major university, and it quickly grew from approximately 150 students in its first offering in 2015 to over 750 students by 2017, making it one of the largest classes at Stanford [6].
CS231n's lecture videos, assignments, and notes were made freely available online, where they reached an audience far beyond Stanford's campus. The course became a de facto entry point into deep learning for an entire generation of AI practitioners and is frequently cited as a formative influence by researchers and engineers working in the field today. The course emphasized building intuition for how neural networks learn, with assignments that required students to implement core components (backpropagation, convolutional layers, batch normalization) from scratch rather than relying on high-level frameworks. Karpathy co-designed the course alongside Fei-Fei Li, and lecture videos have been viewed more than 800,000 times online [3][6].
Karpathy's academic research focused on connecting visual and linguistic understanding. His publication record spans computer vision, video classification, image captioning, and recurrent neural networks. Key papers include:
Large-Scale Video Classification with Convolutional Neural Networks (CVPR 2014). Karpathy, along with George Toderici, Sanketh Shetty, Thomas Leung, Rahul Sukthankar, and Fei-Fei Li, presented one of the first large-scale studies applying CNNs to video classification. The paper introduced the Sports-1M dataset, comprising over 1.1 million YouTube videos across 487 sports categories. The work explored multiple temporal fusion strategies for extending CNNs to process video data and demonstrated significant improvements over feature-based baselines. The paper has received thousands of citations and became a foundational reference in video understanding research [7].
Deep Visual-Semantic Alignments for Generating Image Descriptions (CVPR 2015). This paper presented a model that could generate natural language descriptions of image regions by aligning fragments of sentences with the visual content they describe. Using a joint embedding space and a multimodal RNN decoder, the system achieved strong results on image-sentence retrieval and image captioning benchmarks including Flickr30k and MS COCO [8].
ImageNet Large Scale Visual Recognition Challenge (International Journal of Computer Vision, 2015). Co-authored with Olga Russakovsky, Jia Deng, and others from the ImageNet team, this paper provided a comprehensive description of the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) benchmark. Karpathy contributed analysis comparing human performance with CNN performance on ImageNet classification. His experiments revealed that humans struggled disproportionately with fine-grained recognition categories (such as distinguishing between more than 120 dog breeds), and that overall human error and top CNN error rates were converging. This paper has been cited over 30,000 times and remains one of the most influential papers in the history of computer vision [9].
Visualizing and Understanding Recurrent Neural Networks (2015). This work provided tools for interpreting what recurrent neural networks learn about the structure of text, opening a window into the internal representations of sequence models [10].
DenseCap: Fully Convolutional Localization Networks for Dense Captioning (CVPR 2016). Co-authored with Justin Johnson and Fei-Fei Li, this paper introduced the dense captioning task, which requires a model to both localize and describe salient regions in images using natural language. The proposed Fully Convolutional Localization Network (FCLN) could process an image in a single forward pass, required no external region proposals, and was trained end-to-end. The system was evaluated on the Visual Genome dataset, which contains 94,000 images and 4.1 million region-grounded captions [11].
| Paper | Venue | Year | Key Contribution |
|---|---|---|---|
| Large-Scale Video Classification with CNNs | CVPR | 2014 | Sports-1M dataset; temporal fusion strategies for video CNNs |
| Deep Visual-Semantic Alignments | CVPR | 2015 | Image captioning via joint visual-linguistic embeddings |
| ImageNet Large Scale Visual Recognition Challenge | IJCV | 2015 | Benchmark description; human vs. CNN performance analysis |
| Visualizing and Understanding RNNs | arXiv | 2015 | Interpretability tools for recurrent neural networks |
| DenseCap | CVPR | 2016 | Dense captioning with fully convolutional localization |
In May 2015, Karpathy published "The Unreasonable Effectiveness of Recurrent Neural Networks" on his personal blog. The post became one of the most widely read introductions to RNNs in the deep learning community. It demonstrated how character-level language models trained on various text corpora could learn to generate Shakespeare, Wikipedia articles, LaTeX source code, and Linux kernel code. The post accompanied the release of char-rnn, an open-source implementation of multi-layer LSTM character-level language models written in Torch (Lua). The char-rnn repository allowed users to train character-level models on any text dataset and sample new text that mimicked the style and structure of the training data. The project became one of the early viral open-source deep learning tools, inspiring numerous reimplementations and adaptations [12].
Karpathy was among the founding members of OpenAI when the organization was announced in December 2015. The founding team included Sam Altman, Elon Musk, Ilya Sutskever, Greg Brockman, Wojciech Zaremba, John Schulman, and several others. At OpenAI, Karpathy contributed to research on generative models, reinforcement learning, and the early explorations of language modeling that would eventually lead to the GPT series of models. His time at OpenAI overlapped with a period of foundational research at the organization, when the team was relatively small and focused on publishing open research [1].
In June 2017, Karpathy left OpenAI to join Tesla as Director of AI, reporting directly to Elon Musk. He was subsequently promoted to Senior Director of AI and head of the Autopilot Vision team. At Tesla, his responsibilities encompassed the full stack of the company's autonomous driving AI: neural network architecture design, training infrastructure, data labeling operations, and the deployment of Autopilot and Full Self-Driving (FSD) features to Tesla's global fleet of vehicles [13].
One of Karpathy's most significant and controversial contributions at Tesla was championing the "vision-only" approach to autonomous driving. While most other companies developing self-driving technology (including Waymo, Cruise, and most academic research groups) relied on a combination of cameras, lidar, radar, and high-definition maps, Tesla under Karpathy's technical leadership pursued a strategy that used cameras as the primary sensor, augmented by neural networks that interpreted the visual data directly. In 2021, Tesla went further and removed radar from its newer vehicles, relying entirely on camera-based perception [13].
Karpathy argued that the human visual system demonstrates that cameras alone provide sufficient information for driving, and that the key challenge was building neural networks capable of extracting the necessary information from image data. This approach required building massive data pipelines, which Karpathy referred to as the "data engine": a system for automatically identifying edge cases in the fleet's driving data, labeling them, and using them to retrain and improve the neural networks in a continuous feedback loop [13].
The vision-only approach generated significant debate within the autonomous driving community. Critics argued that relying solely on cameras introduced unnecessary risk, particularly in adverse conditions (rain, fog, glare) where cameras perform poorly. Supporters pointed to the cost advantages and the theoretical sufficiency of visual information. As of 2026, Tesla continues to use the vision-centric approach for its FSD system, though the company has reintroduced radar on some models [13].
Karpathy delivered technical presentations at Tesla's AI Day events in 2021 and 2022, providing unusually detailed looks at the company's neural network architectures, training infrastructure, and data pipeline. These presentations were widely watched and discussed in the AI community, as they offered a rare window into the engineering of a production-scale AI system processing data from millions of vehicles on the road. His explanations of Tesla's multi-camera "BEV" (bird's-eye view) neural network architecture and the auto-labeling pipeline were particularly well received [13].
Karpathy announced his departure from Tesla in July 2022, posting on X (then Twitter) that "it's been a great pleasure to help Tesla towards its goals over the last 5 years and a difficult decision to part ways." He indicated that he wanted to spend time on personal projects, including education and content creation [14].
After leaving Tesla, Karpathy entered a period of independent work focused primarily on education and open-source software. During this time, he produced several notable projects:
minGPT and nanoGPT. Karpathy released minGPT (2022) and its successor nanoGPT (2023), minimal implementations of the GPT architecture in PyTorch. These projects distilled the core ideas behind Transformer-based language models into clean, readable codebases of a few hundred lines. nanoGPT, in particular, became extremely popular on GitHub, earning tens of thousands of stars. It was designed to be simple enough for a single person to understand completely while still being capable of training a functional language model on consumer hardware. The nanoGPT repository could reproduce GPT-2 (124M parameters) on OpenWebText, running on a single 8xA100 40GB node in about four days [15].
YouTube channel. Karpathy launched a YouTube channel focused on deep learning education. His video series "Neural Networks: Zero to Hero" walked viewers through building neural networks from scratch, starting with basic gradient computation and building up to a full implementation of a GPT-style language model. His longer-form videos, including "Deep Dive into LLMs like ChatGPT" (3 hours 31 minutes, released February 2025) and "How I Use LLMs" (2 hours 11 minutes), attracted large audiences. As of early 2026, his YouTube channel has over 1 million subscribers [16].
On February 9, 2023, Karpathy announced that he was returning to OpenAI. His second stint at the company lasted roughly one year. He worked on research projects related to GPT-4 and other efforts, though the specific details of his work during this period were not fully disclosed publicly [1].
Karpathy departed OpenAI for the second time on February 13, 2024. In a post on X, he wrote: "Nothing 'happened' and it's not a result of any particular event, issue or drama. Actually, being at OpenAI over the last ~year has been really great." He described his departure as a natural decision to pursue personal projects and expressed no ill will toward the organization [17].
Since leaving OpenAI in February 2024, Karpathy has focused on building Eureka Labs and expanding his presence as a public educator and thought leader in AI.
On July 16, 2024, Karpathy announced the founding of Eureka Labs, a San Francisco-based AI-native education startup. The company's central premise is that AI can fundamentally transform education by providing every student with the equivalent of a world-class personal tutor. Eureka Labs' first planned course is "LLM101n: Let's Build a Storyteller," an undergraduate-level course in which students build a working language model from scratch, guided by an AI teaching assistant [18].
Karpathy has described his vision for Eureka Labs as creating a new kind of educational institution where the core course content is designed by human domain experts but the delivery, exercises, feedback, and personalization are handled by AI. He sees this as addressing the fundamental bottleneck in education: the scarcity of excellent teachers. The startup plans to run both digital and physical cohorts of students going through the materials together [18].
On February 2, 2025, Karpathy posted on X: "There's a new kind of coding I call 'vibe coding,' where you fully give in to the vibes, embrace exponentials, and forget that the code even exists. It's possible because the LLMs (e.g. Cursor Composer w Sonnet) are getting too good" [2].
The term described an approach to software development in which the programmer provides high-level natural language instructions to an AI coding assistant and accepts the generated code without carefully reviewing or understanding every line. Karpathy noted that he used voice input (via SuperWhisper) to describe what he wanted, accepted all suggestions from the AI, and re-prompted when something looked off rather than manually debugging. He later reflected that the tweet was a casual "shower of thoughts" post that he simply fired off, and that after 17 years on Twitter he still could not predict which posts would go viral [2].
The concept struck a nerve. Within weeks, "vibe coding" had become one of the most discussed terms in the technology world. The post was viewed over 4.5 million times. It inspired debate about the future of software engineering, the role of code literacy, and whether AI-assisted development represented a democratization of programming or a dangerous abdication of understanding. In November 2025, Collins English Dictionary named "vibe coding" its Word of the Year for 2025, defining it as "the use of artificial intelligence prompted by natural language to write computer code" [19].
By early 2026, Karpathy himself noted that the concept had already evolved. As language models grew more capable, what had initially seemed like a casual, experimental approach was becoming a standard professional workflow, though typically with more oversight and scrutiny than the original "vibe coding" philosophy implied. Karpathy suggested that the era of pure vibe coding was already giving way to more structured forms of AI-assisted development [20].
Karpathy's open-source work is distinctive for its pedagogical intent. Rather than building production-ready software, he deliberately creates minimal implementations that sacrifice features for clarity, making it possible for learners to read and understand entire systems.
| Project | Year | Language | Description | GitHub Stars |
|---|---|---|---|---|
| char-rnn | 2015 | Lua (Torch) | Multi-layer LSTM character-level language models | 11k+ |
| micrograd | 2020 | Python | Scalar-valued autograd engine and neural network library (~150 lines) | 15k+ |
| minGPT | 2022 | Python (PyTorch) | Minimal GPT implementation (~300 lines) | 20k+ |
| nanoGPT | 2023 | Python (PyTorch) | Simplified GPT training codebase; reproduces GPT-2 (124M) | 40k+ |
| minbpe | 2024 | Python | Minimal Byte Pair Encoding tokenizer implementation | 12k+ |
| llm.c | 2024 | C/CUDA | GPT-2 training in ~1,000 lines of C with no dependencies | 27k+ |
| nanochat | 2025 | Python (PyTorch) | Full-stack ChatGPT clone pipeline; train for ~$100 in ~4 hours | 8k+ |
| microgpt | 2026 | Python | Single-file 200-line GPT: tokenizer, autograd, training, inference, no dependencies | New |
| autoresearch | 2026 | Python | Autonomous AI agent loop for running ML experiments (~630 lines) | New |
Released in 2020, micrograd implements a complete scalar-valued autograd engine and a small neural network library in roughly 150 lines of Python, making it possible for beginners to understand the entire mechanism of backpropagation by reading a single file. The project deliberately operates at the scalar level rather than using tensors, prioritizing pedagogical clarity over computational efficiency [21].
In 2024, Karpathy released llm.c, an implementation of GPT-2 training in approximately 1,000 lines of C code with no external dependencies beyond the C standard library and CUDA. The project demonstrated that the core computation of training a large language model, while typically wrapped in complex framework code, is fundamentally straightforward. The C implementation matched the output of the equivalent PyTorch reference code, making it clear that the essential algorithm could be expressed without a deep learning framework [22].
Released in early 2024, minbpe provides a minimal, clean implementation of the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization. The repository implements three tokenizer variants and supports training a vocabulary from text, encoding text to tokens, and decoding tokens back to text. Like Karpathy's other projects, minbpe is designed primarily for educational use [23].
Released in October 2025, nanochat is the spiritual successor to nanoGPT. While nanoGPT covered only pretraining, nanochat provides a full-stack pipeline covering tokenization, base pretraining, mid-training on chat and tool-use data, supervised fine-tuning, optional reinforcement learning on GSM8K, evaluation, and serving through both a CLI and a ChatGPT-like web UI. Karpathy described it as a way to build "the best ChatGPT that $100 can buy," with training taking roughly four hours on an 8xH100 GPU node [24].
In March 2026, Karpathy released autoresearch, a 630-line Python script that implements an autonomous AI agent loop for running machine learning experiments. The system gives an AI agent a small LLM training setup and lets it experiment autonomously: modifying code, training for a short period, checking whether results improved, keeping or discarding changes, and repeating the cycle. In one test, the agent processed approximately 700 autonomous changes over two days and found roughly 20 additive improvements that transferred to larger models, reducing the "Time to GPT-2" leaderboard metric from 2.02 hours to 1.80 hours (an 11% efficiency gain). Karpathy described the project as a glimpse of how AI research labs will operate in the future, stating that the goal is "not to emulate a single PhD student" but "to emulate a research community of them" [25].
Karpathy occupies an unusual position in the AI community: he is both a world-class practitioner (having led AI teams at two of the most prominent AI organizations in the world) and one of the field's most effective educators. His ability to distill complex technical concepts into clear explanations, whether in blog posts, YouTube videos, or open-source code, has made him enormously influential.
| Medium | Reach | Notable works |
|---|---|---|
| Stanford CS231n | Thousands of in-person students; millions of online viewers | Lecture videos, assignments, course notes |
| Blog (karpathy.github.io) | Widely read in AI community | "The Unreasonable Effectiveness of Recurrent Neural Networks" (2015) |
| YouTube | 1 million+ subscribers | "Neural Networks: Zero to Hero," "Deep Dive into LLMs like ChatGPT," "How I Use LLMs" |
| GitHub | Tens of thousands of stars across projects | char-rnn, micrograd, minGPT, nanoGPT, minbpe, llm.c, nanochat, microgpt, autoresearch |
| X (Twitter) | ~1.9 million followers | Coined "vibe coding"; regular commentary on AI developments |
Karpathy has articulated an influential framework for understanding the evolution of software through three paradigms. "Software 1.0" refers to traditional code written explicitly by humans. "Software 2.0" is the paradigm in which neural networks are trained on data, with the "code" (the network's weights) being learned rather than written. Karpathy coined this term in a 2017 blog post that generated significant discussion in the AI community. "Software 3.0" extends this to AI systems guided by natural language prompts, where the model itself becomes the runtime and the prompt becomes the program [26].
This framework has been widely adopted in discussions about the future of programming and the role of AI in software development.
| Period | Role | Organization |
|---|---|---|
| 2015-2017 | Founding member, research scientist | OpenAI |
| 2017-2022 | Senior Director of AI, head of Autopilot Vision | Tesla |
| 2022-2023 | Independent educator, open-source developer | Independent |
| 2023-2024 | Researcher | OpenAI (second stint) |
| 2024-present | Founder | Eureka Labs |
| 2024-present | Educator and content creator | Independent (YouTube, GitHub, blog) |
Karpathy has expressed nuanced views on the trajectory of AI. In a 2025 podcast interview, he estimated that artificial general intelligence (AGI) is likely still roughly a decade away, pushing back against more aggressive timelines predicted by some of his peers. He has argued that while current large language models are impressive, they lack key capabilities (robust reasoning, genuine understanding, reliable factual knowledge) that would be needed for AGI [27].
On AI safety, Karpathy has taken a pragmatic rather than alarmist position. He has emphasized the importance of understanding what AI systems are actually doing (interpretability) and building robust engineering practices around AI deployment, rather than focusing primarily on speculative long-term risks [27].
In June 2025, commenting on claims that self-driving technology was a solved problem, Karpathy cautioned against such conclusions, warning that the gap between impressive demos and reliable, fully autonomous operation in all conditions remained substantial [28].
His influence extends well beyond his formal roles. Through his educational content, open-source projects, and public commentary, Karpathy has shaped how an entire generation of developers and researchers thinks about deep learning, language models, and the practice of AI engineering.
Karpathy lives in the San Francisco Bay Area. He is active on X (formerly Twitter) and YouTube, where his posts on AI topics regularly reach millions of people. He has described himself as someone who learns best by building things from scratch, a philosophy that pervades both his educational approach and his personal research projects [1].