Yann Andre LeCun (born July 8, 1960) is a French-American computer scientist whose pioneering work on convolutional neural networks (CNNs) helped lay the groundwork for modern computer vision and deep learning. He shared the 2018 ACM A.M. Turing Award with Geoffrey Hinton and Yoshua Bengio for conceptual and engineering breakthroughs that made deep neural networks a critical component of computing [1]. LeCun served as the Chief AI Scientist at Meta (formerly Facebook) from 2013 until his departure in late 2025, and he has held a professorship at New York University (NYU) since 2003.
LeCun is known not only for his technical contributions but also for his outspoken views on the future of artificial intelligence. He has been one of the most vocal critics of what he considers overblown fears about AI existential risk, arguing that current AI systems are nowhere near the kind of autonomous agency that would pose a threat to humanity. At the same time, he has been a passionate advocate for open-source AI, pushing for the public release of Meta's LLaMA family of large language models. In early 2026, he launched Advanced Machine Intelligence Labs (AMI Labs), a Paris-based startup focused on building world models using his Joint Embedding Predictive Architecture (JEPA), raising over $1 billion in seed funding [2].
Yann LeCun was born on July 8, 1960, in Soisy-sous-Montmorency, a suburb north of Paris, France. His surname, Le Cun, derives from the old Breton form Le Cunff, originating from the region of Guingamp in northern Brittany. "Yann" is the Breton form of "John" [3]. His father was an aeronautical engineer whose interests in electronics and mechanics shaped LeCun's childhood. Growing up, he was a habitual tinkerer, taking things apart and building circuits, a hands-on curiosity that would later inform his approach to building practical AI systems.
LeCun studied at ESIEE Paris (Ecole Superieure d'Ingenieurs en Electronique et Electrotechnique), where he received his Diplome d'Ingenieur in 1983 [4]. He then pursued a PhD in computer science at the Universite Pierre et Marie Curie (UPMC, now part of Sorbonne University) in Paris. His doctoral advisor was Maurice Milgram, who at the time focused on image processing and automata networks. LeCun has noted that Milgram was not working directly on neural networks or machine learning during their time together, but was interested in what was then called "automata networks"; Milgram began working on neural networks after LeCun graduated [5].
LeCun's PhD thesis, "Modeles connexionnistes de l'apprentissage" (Connectionist Learning Models), was completed in 1987. In it, he proposed an early form of the backpropagation learning algorithm for training neural networks, arriving at similar conclusions to those reached independently by David Rumelhart, Geoffrey Hinton, and Ronald Williams in their landmark 1986 paper [6]. The convergence of ideas was striking: researchers on both sides of the Atlantic were discovering the same fundamental principles for training multi-layer networks.
After completing his PhD, LeCun spent a year as a postdoctoral researcher at the University of Toronto, supervised by Geoffrey Hinton. This was 1987 to 1988, a period when Hinton was actively building the research group that would become one of the world's leading centers for neural network research [4]. The postdoc gave LeCun direct exposure to Hinton's work on Boltzmann machines, distributed representations, and the practical challenges of training deep networks. The experience also established a personal and intellectual connection between the two that would persist for decades, even as their views on AI risk diverged sharply in later years.
In 1988, LeCun joined the Adaptive Systems Research Department at AT&T Bell Laboratories in Holmdel, New Jersey, headed by Lawrence D. Jackel [4]. Bell Labs at the time was one of the world's premier industrial research laboratories, with a tradition of fundamental research that had produced the transistor, information theory, Unix, and the C programming language, among other breakthroughs. For LeCun, it was an environment where he could pursue both theoretical ideas and real-world applications.
It was at Bell Labs that LeCun developed the work for which he is best known: convolutional neural networks. In 1989, LeCun and colleagues published "Backpropagation Applied to Handwritten Zip Code Recognition," which described a neural network architecture that used convolutional layers to learn spatial features directly from pixel data [7]. The network was trained using backpropagation to recognize handwritten digits from zip codes provided by the U.S. Postal Service.
This 1989 system was the prototype for what would become known as LeNet. The architecture evolved over the following years, with the most well-known version, LeNet-5, described in a comprehensive 1998 paper by LeCun, Leon Bottou, Yoshua Bengio, and Patrick Haffner titled "Gradient-Based Learning Applied to Document Recognition" [8]. LeNet-5 used a sequence of convolutional layers, subsampling (pooling) layers, and fully connected layers, an architecture template that, with modifications, remains the basis for virtually all modern CNNs.
The practical impact was immediate and concrete. The bank check recognition system that LeCun helped develop at Bell Labs was deployed by NCR Corporation and other companies. By the late 1990s and early 2000s, this system was reading over 10% of all the checks deposited in the United States [8]. It was one of the first large-scale commercial applications of neural networks and demonstrated that these models could work reliably in production systems handling millions of transactions.
| Aspect | Details |
|---|---|
| First CNN paper | 1989: "Backpropagation Applied to Handwritten Zip Code Recognition" |
| Architecture name | LeNet (LeNet-1 in 1989; LeNet-5 in 1998) |
| Application | Handwritten digit recognition (zip codes, bank checks) |
| Co-authors (1998 paper) | Leon Bottou, Yoshua Bengio, Patrick Haffner |
| Commercial deployment | NCR bank check readers; processed >10% of U.S. checks |
| Key architectural features | Convolutional layers, pooling layers, fully connected layers |
In 1990, LeCun, along with John Denker and Sara Solla, published "Optimal Brain Damage," a technique for pruning neural networks by removing weights that contribute least to the network's performance [9]. The method uses second-derivative information (the diagonal of the Hessian matrix) to estimate the importance of each weight and remove those whose deletion would cause the smallest increase in error.
The name was deliberately playful, but the underlying idea was serious and ahead of its time. Neural network pruning has become an active area of research in the 2020s, as practitioners seek to deploy large models on resource-constrained hardware. LeCun's early work on pruning anticipated these concerns by decades.
Between 1996 and 2001, LeCun was one of the main creators of the DjVu image compression technology, alongside Leon Bottou and Patrick Haffner [10]. DjVu was designed for distributing high-resolution scanned documents over the web, at a time when bandwidth was limited and PDF files of scanned documents were impractically large.
DjVu could compress scanned documents at 300 dpi in black-and-white to 5-30 KB per page (3-8 times better than TIFF Group IV compression), and color documents to 30-100 KB per page (5-10 times better than JPEG) [10]. The technology used a segmentation approach that separated text, background, and foreground layers and compressed each with different algorithms optimized for that type of content.
DjVu was spun off into a company and the format gained adoption in digital libraries and document archives, though it never achieved the ubiquity of PDF. The Internet Archive, among other institutions, used DjVu for serving scanned books.
Another notable contribution from LeCun's Bell Labs period was the Graph Transformer Network (GTN), described in the 1998 paper alongside LeNet-5 [8]. GTNs provided a framework for combining multiple trainable modules (including neural networks) into larger systems that could be trained end-to-end using gradient-based methods. The concept anticipated later developments in differentiable programming and end-to-end learning that became central to modern deep learning practice.
When AT&T split in 1996, LeCun moved to AT&T Labs-Research. He later moved to NEC Laboratories America in Princeton, New Jersey, where he continued his work on machine learning and computer vision.
In 2003, LeCun joined the Courant Institute of Mathematical Sciences at New York University as a professor. He later became the founding director of NYU's Center for Data Science and the Silver Professor of Computer Science, Data Science, Neural Science, and Electrical and Computer Engineering [4].
At NYU, LeCun continued to develop and refine his ideas about neural networks, with a particular focus on unsupervised and self-supervised learning methods that he believed were the key to building more capable AI systems. He also worked extensively on energy-based models.
LeCun has been a long-time proponent of energy-based models (EBMs), a framework for machine learning in which a scalar energy function maps configurations of variables to a single number. Lower energy corresponds to more compatible or more likely configurations. Given observed variables, inference consists of finding the values of unknown variables that minimize the energy [11].
EBMs provide a unifying perspective on many machine learning methods, including discriminative models, generative models, and structured prediction. LeCun's work on EBMs included developing methods for contrastive learning, where the model is trained to assign low energy to observed data points and high energy to "negative" examples. This framework influenced later developments in self-supervised learning and contrastive learning that became important for training vision and language models.
In December 2013, Facebook announced that LeCun would lead its newly created AI Research lab, known as FAIR (Facebook AI Research) [12]. The appointment was part of a broader wave of tech companies hiring top AI academics, with Google having acquired Geoffrey Hinton's startup earlier that year and Baidu recruiting Andrew Ng.
LeCun built FAIR into one of the most respected AI research organizations in the world, with labs in New York, Menlo Park, Paris, Montreal, London, and Tel Aviv. Under his leadership, FAIR published influential work across a wide range of topics, including computer vision, natural language processing, reinforcement learning, and robotics. The lab maintained a strong culture of open publication, releasing papers and code publicly, which was somewhat unusual for a corporate research lab at the time.
LeCun was one of the strongest voices inside Meta pushing for the open release of the company's AI models. His advocacy was instrumental in Meta's decision to release the LLaMA family of large language models, starting with LLaMA in February 2023 and continuing with LLaMA 2 in July 2023 and LLaMA 3 in 2024 [13].
His argument for open-source AI was both philosophical and practical. On the philosophical side, LeCun argued that concentrating AI power in the hands of a few companies was dangerous for democracy and individual freedom. He pointed to the history of the internet and open-source software as precedents: open platforms tend to produce better outcomes than closed ones. On the practical side, he argued that open models enable faster research, better customization, and improved safety through community scrutiny [14].
In a widely shared post on X (formerly Twitter), LeCun wrote: "A hugely important commitment to the openness of Meta's AI ecosystem... Llama 3.1 is free, open, and on par with the best proprietary systems. To maximize performance, safety, customizability, and efficiency, AI platforms must be open" [14].
This stance brought him into conflict with researchers and policymakers who argued that releasing powerful AI models openly increased the risk of misuse. LeCun pushed back forcefully, arguing that the benefits of open AI far outweighed the risks and that attempts to restrict AI research would concentrate power in the hands of a few large companies.
| Year | Meta AI release | LeCun's role |
|---|---|---|
| 2013 | FAIR founded | Director and Chief AI Scientist |
| 2023 (Feb) | LLaMA released | Advocated for open release |
| 2023 (Jul) | LLaMA 2 released | Continued open-source advocacy |
| 2024 | LLaMA 3 released | Promoted as open alternative to proprietary models |
| 2025 (Apr) | LLaMA 4 released | Final major release before LeCun's departure |
In November 2025, LeCun left Meta after twelve years as its Chief AI Scientist [15]. The departure followed a period of organizational upheaval at Meta's AI division. After the disappointing reception of the LLaMA 4 model launch in April 2025, Mark Zuckerberg restructured Meta's AI operations. The company invested $14.3 billion in Scale AI and recruited Alexandr Wang as Meta's new Chief AI Officer to lead a new division called Meta Superintelligence Labs [15]. Under this reorganization, LeCun would have reported to Wang rather than directly to Zuckerberg, a change that reportedly contributed to his decision to leave.
LeCun has stated that his departure was also motivated by a fundamental disagreement about the direction of AI research. He believed Meta, like much of the industry, was too focused on scaling large language models, an approach he considered a dead end for achieving genuine machine intelligence.
In 2019, the Association for Computing Machinery announced that LeCun, Hinton, and Bengio would share the 2018 A.M. Turing Award, carrying a $1 million prize funded by Google [1]. The citation recognized the three for "conceptual and engineering breakthroughs that have made deep neural networks a critical component of computing."
The award committee specifically noted LeCun's contributions to convolutional neural networks, which the citation described as having produced "breakthroughs in processing images, video, speech, and audio." His work on backpropagation, convolutional neural networks, and other methods was recognized as having demonstrated the practical advantages of deep networks [1].
The three laureates are sometimes called the "Godfathers of AI" or the "Deep Learning Troika." The award formalized what the AI research community had long recognized: that the work of these three researchers, pursued through decades of skepticism and underfunding, had transformed both computer science and the technology industry.
LeCun has been one of the most prominent skeptics of AI existential risk narratives. His position has put him in direct, often public disagreement with Geoffrey Hinton, Yoshua Bengio, and other researchers who have warned about the dangers of advanced AI systems.
LeCun's skepticism rests on several pillars:
Intelligence does not imply a desire for dominance. LeCun has argued that the scenario where an AI system spontaneously develops a desire to take control is fundamentally flawed. "The first fallacy is that because a system is intelligent, it wants to take control. That's just completely false," he has said [16]. He compares this to the observation that highly intelligent humans do not all seek world domination; intelligence and the drive for power are separate properties.
Current AI lacks genuine understanding. LeCun has been a persistent critic of large language models, arguing that they perform statistical pattern matching on text without achieving genuine understanding of the world. He has described LLMs as "a statistical trick that predicts words but does not understand anything" [2]. He believes that text alone is an insufficient basis for building systems that truly understand causality, physics, and common sense.
AI is engineered, not emergent. LeCun draws an analogy to turbojets: "I can imagine thousands of scenarios where a turbojet goes terribly wrong. Yet we managed to make turbojets insanely reliable before deploying them widely" [17]. His point is that AI systems are designed and built by humans, and safety can be engineered in, just as it is with other complex technologies.
Open AI research improves safety. LeCun argues that restricting AI research and concentrating it in a few companies actually increases risk, because it reduces the number of people who can scrutinize and improve the technology. Open models allow the broader research community to identify problems and develop solutions.
The disagreement between LeCun and his Turing Award co-laureates has played out in public across social media, conferences, and media interviews. In 2024, the three clashed over California's proposed AI safety bill SB 1047. Hinton signed an open letter supporting the bill, while LeCun publicly criticized supporters, arguing they had a "distorted view" of AI's near-term capabilities [18].
LeCun also directly criticized Hinton and Bengio for what he characterized as providing "ammunition to those who are lobbying for a ban on open AI R&D" [19]. He directed similar criticism at CEOs of major AI companies who issued public warnings about existential risk, suggesting that their motivations were partly competitive (regulatory capture that would benefit incumbents at the expense of smaller players and open-source projects).
Despite the heat of these disagreements, LeCun has maintained that he respects Hinton and Bengio as scientists. The debates, while sometimes sharp, reflect genuine differences in how leading researchers assess the trajectory and risks of AI development.
| Topic | LeCun's position | Hinton's position |
|---|---|---|
| Existential risk from AI | Overblown; AI can be engineered safely | Serious concern; 10-20% chance of human extinction in 30 years |
| Open-source AI models | Essential for safety and democracy | Potentially dangerous; powerful models should be restricted |
| LLMs as path to AGI | Dead end; statistical tricks that lack understanding | Closer to general intelligence than many think |
| AI regulation (e.g., SB 1047) | Opposes; risks stifling research and open-source | Supports; necessary to manage risks |
| Timeline to superhuman AI | Decades away; requires fundamental breakthroughs | Possibly within 5-20 years |
LeCun has argued for years that the path to human-level AI runs not through larger language models but through systems that build internal models of how the world works, what he calls "world models." This vision crystallized in a position paper he published in 2022 titled "A Path Towards Autonomous Machine Intelligence," which laid out his blueprint for an architecture capable of learning, reasoning, and planning in the physical world [20].
The centerpiece of this architecture is the Joint Embedding Predictive Architecture (JEPA). Unlike generative models that learn to predict raw data (such as the next pixel in a video or the next word in a sentence), JEPA learns to predict representations of data in an abstract embedding space [20]. The distinction matters because predicting raw data forces a model to account for every detail (the exact color of every pixel, the precise wording of a sentence), much of which is irrelevant noise. Predicting in abstract representation space allows the model to focus on the underlying structure and dynamics.
In LeCun's framework, a world model trained with JEPA would learn from video, audio, and physical sensor data to understand how objects move, how gravity works, what happens when you push something off a table. This kind of physical intuition, which humans develop in infancy, is something that no current AI system possesses, and LeCun argues it cannot be learned from text alone.
Meta's FAIR lab, under LeCun's direction, published several implementations of the JEPA concept:
I-JEPA (Image-based JEPA), published in 2023, demonstrated that a model trained to predict masked image regions in representation space (rather than pixel space) could learn useful visual features without any labeled data. It outperformed pixel-prediction methods on several downstream tasks while being more computationally efficient [21].
V-JEPA (Video-based JEPA), published in 2024, extended the approach to video, learning to predict masked spatiotemporal regions. The model learned representations that captured physical dynamics, object permanence, and other properties of the visual world [21].
These systems represented steps toward LeCun's vision of world models but were still far from the complete autonomous intelligence architecture he described in his 2022 paper.
In March 2026, LeCun's new Paris-based startup, Advanced Machine Intelligence Labs (AMI Labs), emerged from stealth with $1.03 billion in seed funding, the largest seed round ever raised by a European company [2]. The company was co-founded with Alexandre LeBrun, who serves as CEO, while LeCun holds the title of Executive Chair.
The funding round, announced on March 10, 2026, was co-led by Cathay Innovation, Greycroft, Hiro Capital, HV Capital, and Bezos Expeditions, resulting in a pre-money valuation of $3.5 billion. Other investors included NVIDIA, Eric Schmidt, Jeff Bezos, and major European and Asian funds [2].
AMI Labs represents a direct bet against the prevailing approach of the AI industry. While companies like OpenAI, Anthropic, and Google continue to invest billions in scaling large language models, LeCun is wagering that the future belongs to world models trained on video, audio, and physical sensor data using the JEPA architecture. The company's mission is to build AI systems that learn from reality, not just from language.
| Aspect | Details |
|---|---|
| Company name | Advanced Machine Intelligence Labs (AMI Labs) |
| Headquarters | Paris, France |
| Founded | Late 2025 / Early 2026 |
| Seed funding | $1.03 billion |
| Pre-money valuation | $3.5 billion |
| CEO | Alexandre LeBrun |
| Executive Chair | Yann LeCun |
| Core technology | JEPA-based world models |
| Key investors | Cathay Innovation, Bezos Expeditions, NVIDIA, Eric Schmidt |
The choice to base the company in Paris rather than Silicon Valley was deliberate. LeCun has long been an advocate for European AI research and has maintained strong connections to the French AI community. The decision also reflects the growing competitiveness of Paris as a hub for AI startups and research.
| Contribution | Year | Significance |
|---|---|---|
| Backpropagation (early formulation) | 1987 | PhD thesis independently arrived at backpropagation for neural networks |
| Convolutional neural networks (LeNet) | 1989-1998 | Created the architecture that became the foundation of modern computer vision |
| Optimal Brain Damage | 1990 | Early neural network pruning technique; anticipated modern model compression |
| DjVu image compression | 1996-2001 | High-compression format for scanned documents |
| Graph Transformer Networks | 1998 | Framework for end-to-end trainable modular systems |
| Energy-based models | 2000s-2010s | Unified framework for discriminative and generative learning |
| FAIR founding | 2013 | Built one of the world's top AI research labs |
| Open-source AI advocacy | 2023-2025 | Instrumental in Meta's release of LLaMA models |
| JEPA architecture | 2022-2026 | Proposed alternative path to machine intelligence based on world models |
LeCun has held an unusually wide range of positions across academia and industry:
| Period | Position | Institution |
|---|---|---|
| 1988-1996 | Researcher, Adaptive Systems Research Department | AT&T Bell Labs |
| 1996-2003 | Researcher | AT&T Labs-Research, then NEC Labs America |
| 2003-present | Silver Professor | NYU Courant Institute |
| 2013-2025 | VP and Chief AI Scientist | Meta (formerly Facebook) |
| 2016 | Inaugural chair, Computer Sciences and Digital Technologies | College de France |
| 2026-present | Executive Chair | AMI Labs |
He is a member of the U.S. National Academy of Engineering and the U.S. National Academy of Sciences. In 2025, he was one of seven recipients of the Queen Elizabeth Prize for Engineering, alongside Geoffrey Hinton, Yoshua Bengio, John Hopfield, Bill Dally, Jensen Huang, and Fei-Fei Li [22].
LeCun has three sons. His brother works at Google. He acquired American citizenship after his move to the United States in 1988 and has lived in the New York area for most of his professional career, while maintaining close ties to France [3].
He is known for his prolific and sometimes combative presence on social media, particularly on X (formerly Twitter), where he regularly engages in public debates about AI research, policy, and philosophy. His willingness to argue publicly with other prominent researchers, including his Turing Award co-recipients, has made him one of the most recognizable figures in the AI community. Colleagues describe him as intellectually generous in person, always willing to discuss ideas and debate technical points, even if the public exchanges can seem heated.
LeCun's convolutional neural network is one of the most consequential inventions in the history of computer science. Nearly every modern system for image recognition, video analysis, medical imaging, autonomous driving, and facial recognition uses CNN architectures that trace their lineage directly to LeNet. When AlexNet, developed by Hinton's students Krizhevsky and Sutskever, won the ImageNet competition in 2012, it used a deeper and larger version of the same basic convolutional architecture that LeCun had pioneered at Bell Labs two decades earlier.
Beyond CNNs, LeCun's insistence on the importance of self-supervised learning and his work on energy-based models have influenced the direction of AI research. His JEPA framework, now being developed at AMI Labs, represents an ambitious attempt to move beyond the language model paradigm and build AI systems that understand the physical world.
LeCun's advocacy for open-source AI has had a direct impact on the industry landscape. The release of LLaMA gave researchers, startups, and smaller companies access to powerful language models that would otherwise have been available only to the largest tech companies. This democratization of AI tools has enabled a wave of fine-tuned and specialized models built on top of the LLaMA architecture.
His career also illustrates the productive tension between academic research and industrial application. LeCun has always maintained a foot in both worlds, publishing academic papers while building systems that process millions of real-world transactions. The check-reading system at Bell Labs, the AI infrastructure at Meta, and now AMI Labs all reflect his belief that the best way to advance AI is to build things that work in practice.
Whether the JEPA-based world model approach will prove correct, or whether scaled language models will continue to dominate, is one of the central questions in AI as of 2026. LeCun has staked his reputation and over a billion dollars of investor money on the answer.