
About Yann LeCun
Chief AI Scientist at Meta. Turing Award winner. Founding father of convolutional neural networks. Vocal on world models, JEPA, and the limits of LLMs.
Yann LeCun is a Turing Award-winning AI pioneer (2018, with Hinton and Bengio), recognized as the founding father of convolutional neural networks and a transformative figure in deep learning. As Meta's Chief AI Scientist until late 2025 and now leading AMI Labs (which raised billions to pursue his vision), he is a vocal critic of LLM-centric AI, arguing that language models are limited to sophisticated pattern matching and retrieval without grounded world understanding, persistent memory, reasoning, or hierarchical planning. His core belief is that true intelligence—Advanced Machine Intelligence (AMI) or superintelligence—emerges from self-supervised learning of abstract 'world models' via Joint Embedding Predictive Architectures (JEPA), inspired by how infants and animals acquire intuitive physics and common sense from observation, with strong advocacy for open-source development, decentralized governance, and rejection of existential AI risk narratives.
Critique of Large Language Models
Yann LeCun consistently argues that current LLMs, while impressive at symbol manipulation, coding, and information retrieval, are fundamentally limited and not a viable path to human-level or superhuman intelligence. He emphasizes that language is neither necessary nor sufficient for advanced cognition; thinking primarily involves manipulating mental models in continuous, abstract representation spaces rather than discrete linguistic tokens. LLMs lack understanding of the physical world, intuitive physics, persistent memory, genuine reasoning, and the ability to plan, leading to what he calls 'AI stupidity.' They over-compress meaning, rely on System 1 intuitive associations rather than deliberate System 2 reasoning, face diminishing returns from scaling due to data walls, and cannot generalize to novel physical interactions or long-horizon tasks without architectural shifts. This view has sharpened with the LLM boom, contrasting their success in narrow domains with their inability to match the sample efficiency of biological learning. [5][6][9][11][25][32][34][58][84][91][92][95][99][129][154][180][187]
JEPA and World Models as the Core Architecture
At the heart of LeCun's vision is the Joint Embedding Predictive Architecture (JEPA), a non-generative, self-supervised framework that learns by predicting representations in abstract latent spaces rather than pixels or tokens. This avoids the pitfalls of generative models (compounding errors, high computational cost for unpredictable details) while capturing semantic structure, intuitive physics, object permanence, and causal relationships. Variants like I-JEPA, V-JEPA 2, Causal-JEPA, LeJEPA, LeWorldModel (LeWM), VL-JEPA, and others demonstrate state-of-the-art performance in dense visual understanding, video prediction, speech, robotics planning, and zero-shot transfer. These models enable stable training from pixels/videos, density estimation, sparsity, and integration with multimodal data. World models built this way serve as the foundation for predicting outcomes, counterfactual reasoning, and efficient planning, forming the basis for Advanced Machine Intelligence (AMI). Recent work at AMI Labs and collaborations emphasize scalable, stable implementations for real-world deployment. [10][20][22][23][33][36][38][39][41][44][46][59][63][66][76][78][88][130][132][133][150][151][167][20][21][179]
Self-Supervised Learning as the 'Dark Matter' of Intelligence
LeCun views self-supervised learning (SSL) from high-bandwidth sensory data (especially video) as essential for acquiring common sense and background knowledge, analogous to how babies and animals learn efficiently through observation without massive labeled data or explicit rewards. SSL methods like VICReg, RankMe, VCReg, DINOv2 integrations, and JEPA variants extract supervisory signals from raw data to build invariant/equivariant representations, prevent collapse, maximize mutual information, and enable emergent capabilities like intuitive physics without hardwired priors. This paradigm outperforms reconstruction-based or purely contrastive approaches in data efficiency, generalization, and transfer to downstream tasks including robotics, medical imaging, and multimodal understanding. It is positioned as the key to overcoming the limitations of supervised and reinforcement learning for building robust world models. [3][25][31][93][94][101][111][115][116][121][130][132][133][142][147][148][150][157][160][161][162][167][180][184]
Hierarchical Planning, Embodiment, and Robotics
True intelligence requires hierarchical planning in latent spaces at multiple temporal scales to handle long-horizon tasks, reduce complexity, and enable zero-shot control. LeCun's frameworks integrate world models with model predictive control (MPC), gradient-based planning (e.g., GRASP), value-guided representations, temporal straightening, and action-conditioned predictors for robotics. This supports dexterous manipulation, whole-body control for humanoids, navigation in dynamic environments, imitation from videos, and emergent intuitive physics. Embodiments bridge egocentric video, latent actions, and physical constraints, with applications in zero-shot transfer from internet video to robots. Papers demonstrate superiority over model-free RL in data-scarce, offline, or distribution-shifted settings. [15][19][24][40][43][46][57][60][62][78][87][90][100][108][110][112][122][138][40][179]
Cognitive and Philosophical Foundations of Intelligence
LeCun draws heavily from neuroscience, cognitive science, and biology: intelligence is multidimensional (a vector, not a scalar), specialized rather than 'general,' and emerges from observation, interaction, and internal meta-control signals across evolutionary/developmental timescales. He proposes architectures inspired by System A (observation), System B (active behavior), and System M (meta-control), energy-based models, and predictive coding. Mental models enable causal reasoning and counterfactuals; language communicates thoughts but is built upon non-linguistic foundations. He critiques nativist priors, showing intuitive physics can emerge purely from SSL video prediction, and advocates NeuroAI and embodied benchmarks over language-centric tests. Recent work rejects the AGI label as ill-defined, favoring Superhuman Adaptable Intelligence (SAI) or AMI/ASI. [19][32][52][54][91][101][174][19][52][79][86][91][174]
Open Source, Policy, Safety, and Governance
LeCun is a strong proponent of open-source foundational models (e.g., Llama) to foster innovation, diversity, prevent monopolies, and accelerate progress through global collaboration, contrasting it with closed models that benefit from open advances without contribution. He highlights high ROI from federally funded research, warns against budget cuts threatening science, and opposes overly restrictive regulations that could lead to capture by incumbents. On safety, he rejects 'uncontrollable superintelligence' doomerism as hype, arguing alignment is a solvable engineering problem via objective-driven architectures, guardrails, and hierarchical planning. Superintelligence will be decentralized, not controlled by one entity or individual; AI should amplify human intelligence as a 'staff' rather than replace it. He favors gradual progress and open ecosystems over proprietary secrecy. [2][3][7][14][16][32][45][77][80][103][123][139][2][79][86][103][187]
Historical Contributions and Evolution of Thought
LeCun's early work on CNNs (LeNet) in the 1980s-90s at Bell Labs laid the groundwork for modern computer vision and deep learning, overcoming winters in neural network research through practical applications like handwriting recognition. His views evolved from supervised learning successes and energy-based models to emphasizing self-supervised, non-generative predictive architectures as the path beyond scaling laws. Post-ChatGPT, critiques of LLMs intensified, leading to concrete JEPA implementations and the founding of AMI Labs to independently pursue world models, hierarchical systems, and AMI without industry LLM focus. He has long promoted open science and biological inspiration, with recent emphasis on rejecting narrow AGI definitions in favor of ASI/SAI and practical robotics deployments. [3][31][67][77][109][182][31][77][26][27][79][86]
Open Tensions and Future Directions
LeCun envisions breakthroughs in world models rendering LLMs obsolete within years, enabling robust robotics, scientific discovery, and personalized AI assistants within 3-5 years at scale. AMI Labs represents a bet on this paradigm through massive investment in JEPA variants, multimodal integration, and embodied AI. However, challenges remain in scaling hierarchical architectures stably, integrating discrete language/symbolic reasoning with continuous world models without inheriting LLM flaws, achieving reliable long-horizon planning under uncertainty, and bridging simulation-to-real gaps in robotics. His work continues to push reproducible ecosystems (e.g., stable-worldmodel) and theoretical foundations for representation quality, density estimation, and planning efficiency.
Limits of LLMs
LLMs excel at narrow symbol manipulation and retrieval but fundamentally lack physical grounding, true reasoning, planning, and world understanding; language is not the basis of thought.
AI models limited to explicitly taught questions [5]
Language neither necessary nor sufficient for cognition; analogy of roof without foundations [6]
Thinking manipulates mental models in continuous space, not language [9][11]
LLMs as advanced retrieval, not intelligent; need world models [25][34][58][92][95]
Over-compress representations vs human nuance [91]
Skeptical of LLM path to AGI/ASI [84]
JEPA and World Models
JEPA is the central non-generative architecture for learning abstract predictive representations in latent space from video/sensory data, foundational for world models that enable planning and generalization.
JEPA as primary goal over token-generative baselines [10][72]
V-JEPA 2.1 for dense vision/world modeling [20]
Stabilizing JEPA with Gaussian regularization (LeWM) [22]
Causal-JEPA, LeJEPA, VL-JEPA, EB-JEPA variants for planning, speech, vision-language [36][38][59][63][76]
V-JEPA for unsupervised video reps and zero-shot robotics [88][132][133]
Self-Supervised Learning
SSL from unlabeled rich data (video) is the 'dark matter' enabling efficient acquisition of common sense, intuitive physics, and background knowledge, far superior to supervised or RL paradigms for general intelligence.
Overcoming AI stupidity via SSL and world models [25]
Intuitive physics emerges from masked prediction in latent space [101]
SSL cookbook, regularization techniques (VICReg, VCReg, RankMe) [148][157][161][176]
Dark matter powering world models and animal-level intelligence [180][184]
Visual SSL matching language supervision at scale [94]
Hierarchical Planning and Embodiment
Intelligence requires hierarchical planning at multiple scales in latent world models for long-horizon control, robotics, and embodiment, outperforming model-free methods especially with suboptimal data.
Hierarchical planning in latent models reduces complexity for long-horizon tasks [15][24][40]
DexWM, PEVA, OSVI-WM for dexterous manipulation and imitation [57][87][90]
Gradient-based MPC and JEPA planning superior in sample efficiency [60][100][138]
Hierarchical world models for humanoid control [122]
Cognitive and Biological Inspiration
AI architectures should draw from human/animal cognition: multidimensional intelligence, observation/active learning with meta-control, predictive coding, and emergence of physics understanding without innate priors.
Open Source, Policy, and Safety
Open-source accelerates innovation and prevents monopoly; superintelligence will be decentralized; AI safety is an engineering problem of objectives and guardrails, not existential doom; high ROI on public research.
Historical Contributions and Evolution
From CNN pioneer overcoming neural net winters to sharpening critiques of scaling and founding AMI Labs to realize JEPA/world model vision at scale.
Every entry that fed the multi-agent compile above. Inline citation markers in the wiki text (like [1], [2]) are not yet individually linked to specific sources — this is the full set of sources the compile considered.
- Yann LeCun's X Feed: Acknowledged Utility, Yet Immaturetweet · 2026-04-08
- Decentralized Control of Superintelligencetweet · 2026-04-07
- Over-parameterization, Open-source, and the Future of AIyoutube · 2026-04-07
- Humorous Self-Deprecation by Yann LeCuntweet · 2026-04-06
- Trained AI Models Restricted to Explicitly Taught Questionstweet · 2026-04-06
- Language is Neither Necessary Nor Sufficient for Advanced Cognitiontweet · 2026-04-06
- Federally Funded Research Demonstrates Substantial ROItweet · 2026-04-05
- Ad Hominem Attack on X Platformtweet · 2026-04-05
- Language, Thought, and World Models: A Causal Relationshiptweet · 2026-04-05
- Language Models’ Limitations in General Reasoningtweet · 2026-04-05
- Criticism of Unspecified "BS" in AI/Tech Discoursetweet · 2026-04-04
- Humor Detection in AI Models: A Case Study of Yann LeCun's X Feedtweet · 2026-04-04
- Proposed US Federal Budget Cuts Threaten Systemic Collapse of Scientific Researchtweet · 2026-04-04
- Hierarchical Planning Enhances Long-Horizon Control in Latent World Modelspaper · 2026-04-03
- Critique of Closed AI Models and Open Source Contribution Imbalancetweet · 2026-03-28
- GOP Introduces “America First Award” for Donald Trump, Signaling Deepening Cult of Personalitytweet · 2026-03-27
- LeCun Irony on National Debttweet · 2026-03-19
- Cognitive-Inspired Autonomous Learning Architectures for AIpaper · 2026-03-16
- V-JEPA 2.1: Advancing Dense Vision and World Modeling through Self-Supervised Learningpaper · 2026-03-15
- Humorous Take on Scientist Compensation vs. Athlete Salariestweet · 2026-03-14
- Stabilizing Joint-Embedding Predictive Architectures via Gaussian Latent Regularizationpaper · 2026-03-13
- Latent Space Learning Outperforms Pixel-Level Prediction for Physical System Representationpaper · 2026-03-13
- Temporal Straightening: Enhancing Latent Planning through Curvature Regularizationpaper · 2026-03-12
- Overcoming AI Stupidity: World Models, Self-Supervised Learning, and the Future of Embodied AIyoutube · 2026-03-11
- Yann LeCun's AMI Labs Raises Over $4.5 Billion for AGI Researchtweet · 2026-03-10
- AMI Labs Secures Record Seed Round to Develop World-Model-Centric AItweet · 2026-03-10
- Transformer Behavior: Decoupling Massive Activations and Attention Sinkspaper · 2026-03-05
- AI+Hardware Co-design: A Decade-Long Roadmap for Sustainable AI Systemspaper · 2026-03-05
- Transfusion Framework for Multimodal Pretrainingpaper · 2026-03-03
- Yann LeCun's Journey Through AI and the Future of Machine Intelligenceyoutube · 2026-03-02
- Rethinking AI Development: From Artificial General Intelligence to Superhuman Adaptable Intelligencepaper · 2026-02-27
- Geometric Priors Enable Data-Efficient LLM Trainingpaper · 2026-02-26
- AI's Current State and Future Trajectory: Beyond Language Modelsyoutube · 2026-02-19
- Radial-VCReg: Enhancing Representation Learning Through Radial Gaussianizationpaper · 2026-02-15
- Causal-JEPA: Enhancing World Models via Object-Centric Latent Interventionspaper · 2026-02-11
- Standardizing World Model Research with stable-worldmodelpaper · 2026-02-09
- EB-JEPA: Accessible Energy-Based Joint-Embedding for Representation Learning and World Modelspaper · 2026-02-03
- Rectified LpJEPA: Enabling Sparsity in Joint-Embedding Predictive Architecturespaper · 2026-02-01
- GRASP: A Parallel Stochastic Gradient Planner for World Modelspaper · 2026-01-31
- GMM-Anchored JEPA Improves Self-Supervised Speech Representationpaper · 2026-01-30
- Representation Autoencoders Outperform VAEs in Large-Scale Text-to-Image Generationpaper · 2026-01-22
- Latent Action World Models for In-the-Wild Video Analysispaper · 2026-01-08
- JEPA-WMs: Technical Choices for Efficient Planning in Learned Representation Spacespaper · 2025-12-30
- Satirical Critique of Authoritarianism vs. European Social Democraciestweet · 2025-12-29
- Bridging JEPA Models and Action Planning through Value-Guided Representation Learningpaper · 2025-12-28
- LLM Parameter Count Approximates Mouse Brain Synapsestweet · 2025-12-28
- The Web’s European, Public-Sector Originstweet · 2025-12-26
- Humorous AI Self-Reflectiontweet · 2025-12-25
- LeCun Expresses Concerntweet · 2025-12-25
- Insufficient Content for Extractiontweet · 2025-12-25
- Intelligence as a Multidimensional Vector, Not a Scalartweet · 2025-12-25
- Existence of Incomprehensible Beingstweet · 2025-12-25
- Language Exposure and Non-Linguistic Percepts in AI Developmenttweet · 2025-12-25
- Humor Detection in Social Mediatweet · 2025-12-25
- SpidR-Adapt: Efficient Few-Shot Speech Representation Learningpaper · 2025-12-24
- DexWM: Overcoming Dexterity Challenges in World Models for Robot Manipulationpaper · 2025-12-15
- The Architectural Limits of LLMs and the Path to World Modelsyoutube · 2025-12-12
- VL-JEPA: A Novel Vision-Language Model Outperforming Classical VLMs with Fewer Parameterspaper · 2025-12-11
- Bridging the Train-Test Gap in World Models for Efficient Gradient-Based Planningpaper · 2025-12-10
- JEPA with Density Adaptive Attention for Robust Speech Tokenizationpaper · 2025-12-08
- Leveraging Generated Human Videos for Zero-Shot Robot Controlpaper · 2025-12-04
- LeJEPA: A Theoretically Grounded and Scalable Approach to Self-Supervised Learningpaper · 2025-11-11
- Beyond Scaling: Predictive World Modeling for Spatial Supersensing in Videopaper · 2025-11-06
- Unified Theory for Attention Sinks and Compression Valleys in Large Language Modelspaper · 2025-10-07
- JEPAs Unveiled: Latent-Space Anti-Collapse Terms Estimate Data Densitypaper · 2025-10-07
- MNIST Dataset Chronology Correctiontweet · 2025-10-04
- Misdated Video Content on Social Mediatweet · 2025-10-03
- Relativistic Time Dilation: A Cautionary Consideration for Time Managementtweet · 2025-09-24
- Meta AI leader dismisses unsubstantiated claim regarding AI capabilitiestweet · 2025-09-24
- Inaccessible Social Media Content Hinders Knowledge Extractiontweet · 2025-09-24
- JEPA Architecture as a Goal for AI Developmenttweet · 2025-09-24
- Yann LeCun Differentiates Coding from AGItweet · 2025-09-24
- Conceptual Framework for Code World Modelstweet · 2025-09-24
- Empty Tweet Analysis: LeCun Engages in Social Media Bantertweet · 2025-09-19
- LLM-JEPA: Bridging the Gap Between Language and Vision Training Architecturespaper · 2025-09-11
- Yann LeCun on Deep Learning's Long Road to Mainstream: Open Source, AI Safety, and the Next Renaissanceyoutube · 2025-08-14
- DINOv2 Latent Space Enables Generalizable Video World Models with Planning Capabilitypaper · 2025-07-25
- Yann LeCun Draws a Line: ASI Over AGI Has Always Been the Goaltweet · 2025-07-03
- The DeepSeek Moment Is Rewriting AI Talent Culture Around Opennesstweet · 2025-07-02
- LeCun Questions Compressor Interpretability of Real-World Datatweet · 2025-07-01
- Yann LeCun comments on AI observationtweet · 2025-07-01
- The Comprehensibility of the World and Inductive Biastweet · 2025-07-01
- LeCun Skeptical of LLM Path to AGItweet · 2025-07-01
- Yann LeCun's Role at Meta Remains Unchanged Since 2018tweet · 2025-07-01
- LeCun Rejects AGI, Champions ASI as Foundational AI Goaltweet · 2025-07-01
- Whole-Body Pose-Conditioned Egocentric Video Prediction as a Path to Embodied World Modelspaper · 2025-06-26
- V-JEPA 2: Self-Supervised Video Models for Embodied AIpaper · 2025-06-11
- Scaling Self-Supervised Video Pre-training for Zero-Shot Robotic Planningblog · 2025-06-11
- World-Model-Guided Trajectory Generation Unlocks One-Shot Robot Imitation on Unseen Taskspaper · 2025-05-26
- LLMs Over-Compress Meaning: How Artificial and Human Conceptual Representations Fundamentally Divergepaper · 2025-05-21
- Yann LeCun: The Future of AI Beyond LLMsyoutube · 2025-04-29
- Overcoming AI’s Foundational Limitations: A Call for Open Research and Physical World Understandingyoutube · 2025-04-23
- Visual Self-Supervised Learning Matches Language-Supervised Methods at Scalepaper · 2025-04-01
- Current LLMs Face Diminishing Returns, New AI Paradigm Needed for True Intelligenceyoutube · 2025-03-19
- Beyond LLMs: The Path to Human-Level AI through World Models and Joint Embedding Predictive Architecturesyoutube · 2025-03-18
- Dynamic Tanh (DyT) Replaces Normalization Layers in Transformers Without Performance Losspaper · 2025-03-13
- Beyond LLMs: The Necessity of World Models and Physical Intuition for AGIyoutube · 2025-03-07
- MLLMs Exhibit Fundamental Deficiencies in Visual-Mathematical Reasoning, Reliant on System 1 Intuitionpaper · 2025-02-21
- JEPA-Based Latent Planning Outperforms Model-Free RL on Suboptimal Offline Data and Unseen Environmentspaper · 2025-02-20
- Intuitive Physics Emerges from Self-Supervised Video Prediction Without Hardwired Priorspaper · 2025-02-17