absorb.md

About François Chollet

Creator of Keras. Now at Anthropic. Author of "Deep Learning with Python". Most thoughtful voice on AGI benchmarking via the ARC challenge.

François Chollet is the creator of Keras, author of 'Deep Learning with Python,' and a leading critic of LLM scaling hype, now at Anthropic pushing symbolic program synthesis for AGI. He defines intelligence as efficient skill acquisition in novel situations, benchmarked via his ARC challenge where LLMs score near-zero while humans excel. Chollet advocates hybrid AI combining deep learning's perception with discrete reasoning for true generalization, rejecting exponential 'intelligence explosion' narratives.

Biography and Background

François Chollet created Keras, a high-level deep learning library [8][28][40], authored Deep Learning with Python with companion notebooks [28], and founded the ARC-AGI benchmark to measure fluid intelligence [1][32][33]. Now at Anthropic, he critiques DL myopia [3] and leads efforts in symbolic learning [22].

Keras and Deep Learning Tools

Chollet champions Keras with JAX for optimal AI development [8][12][13][14], introducing Keras Kinetic for cloud TPU/GPU execution [13][14]. He provides extensive code resources: transfer learning tutorials [55][56], data augmentation CNNs [57], functional API designs [58], and model repos [40][46]. Early Gists cover RNN optimizations [52], Nelder-Mead [44], and backend integrations [49][50][51][59].

Critique of Deep Learning and LLMs

DL excels at interpolative perception but fails discrete reasoning and generalization [1][10][29][30][34][36][37][43][45]. LLMs are 'vector program databases' for memorization, lacking fluid intelligence [5][6][29][36][37][42]; they score <1% on ARC vs. human 97-98% [26][29][33][36]. Curve-fitting limits complex systems [9][10]; symbolic methods outperform [4][11].

AGI Definition and Benchmarks

AGI is human-level learning efficiency on novel tasks, not benchmark scores [25][47]. ARC-AGI series [1][2][23][24][26][27][32][33][35][38] tests core knowledge priors and program synthesis; ARC-3 adds interactivity [26][27]. Roadmap: annual unsaturated releases [24]; ARC Prize incentivizes progress [33][38].

Symbolic Learning and Program Synthesis

Chollet pushes 'program synthesis' over scaling [1][4][22][30][34][36][38][45], using DL for intuition + search for reasoning [30][36]. Symbolic compression enables extreme generalization [11]; NDIA lab targets hybrid systems [22][30]. Active inference aids LLMs on ARC [39].

Intelligence Theory

Intelligence is skill acquisition efficiency, not skill [20][25][47]; bounded optimality ratio [20]. Humans near-optimal with tools [19][20]; AGI shifts class divides to cognitive agency [21]. Rejects intelligence explosion [48][60]; progress linear despite resources [60].

Societal Impacts and Broader Views

AI favors established firms [16][17]; enterprise AI needs security [18]. Critiques web for 'collective stupidity' [61][62]; calls for creativity-focused platforms [61]. Scientific progress linear [60][11].

LLM Limitations

LLMs memorize patterns but lack reasoning, generalization, and fluid intelligence; ARC exposes this gap.

  • LLMs are 'vector program databases' scoring near-zero on ARC [29][36]

  • Base LLMs lack fluid intelligence, LRMs show promise [5][6]

ARC-AGI Benchmark

Core tool for measuring AGI via novel task efficiency; annual evolution to stay unsaturated.

  • ARC-AGI-1/2/3 test program synthesis and interactivity [26][32][33]

  • Roadmap for ARC-4 in 2027 [24]

Symbolic Program Synthesis

Path to AGI: hybrid DL perception + discrete search for optimal generalization.

  • Symbolic learning reverse-engineers programs [1][4][22]

  • NDIA lab builds DL-guided synthesis [22][30]

Intelligence as Efficiency

True intelligence is rapid skill acquisition in novelty, not memorized performance.

  • AGI = human-level learning efficiency [25][47]

  • Bounded conversion ratio [20]

Keras Ecosystem

Advocates Keras/JAX/Kinetic for efficient ML; extensive tutorials and tools.

Linear Progress, No Singularity

Exponential resources yield linear gains due to rising discovery difficulty.

  • Scientific progress linear [60]

  • Rejects intelligence explosion [48]

DL Strengths and Limits

DL great for interpolation/perception, fails discrete reasoning.

tool · 64 mentions
tool · 35 mentions
tool · by Greg Brockman · 30 mentions
tool · by Simon Willison · 18 mentions
tool · 13 mentions
twitter
tool · 12 mentions
app · by Casey Newton · 11 mentions
repo · by Garry Tan · 9 mentions
movie · by Andy Weir · 8 mentions
book · 7 mentions
tool · by Google · 7 mentions
tool · 6 mentions
app · 6 mentions
paper · by Francois Chollet · 5 mentions
kaggle
service · by Google · 4 mentions
app · 4 mentions
book · by Andy Weir · 3 mentions
masterclass
service · 3 mentions
tool · 3 mentions
book · by François Chollet · 3 mentions

Every entry that fed the multi-agent compile above. Inline citation markers in the wiki text (like [1], [2]) are not yet individually linked to specific sources — this is the full set of sources the compile considered.

  1. Emerging AI Inference Accelerators: A Landscape of Specializationyoutube · 2026-04-10
  2. AI Collapses Coordination Costs to One Person, Enabling the Solo-Run Conglomerateyoutube · 2026-04-10
  3. Scaling Agentic Coding Requires Organizational Infrastructure, Not Just Better Modelsyoutube · 2026-04-10
  4. AI Compute Economy Matures: Token Scarcity, Meta's Strategic Dilemma, and the Rise of Financial Infrastructure for GPU Marketsyoutube · 2026-04-10
  5. Symbolic Descent: The Case for Replacing Deep Learning's Parametric Foundation with Minimal Symbolic Modelsyoutube · 2026-04-10
  6. François Chollet's Case Against Scaling: Why AGI Requires Skill Acquisition, Not Task Automationyoutube · 2026-04-10
  7. Rethinking AGI Development: Beyond Deep Learning Limitationsyoutube · 2026-04-09
  8. ARC Prize Foundation Seeks Benchmark Lead for AGI Developmenttweet · 2026-04-07
  9. Critique of Deep Learning Research Myopiatweet · 2026-04-07
  10. Symbolic Learning for Generative Program Reverse Engineeringtweet · 2026-04-06
  11. Emergent Reasoning in Advanced Language Modelstweet · 2026-04-06
  12. LLMs Lack Fluid Intelligence, While LRMs Show Promise in Reasoningtweet · 2026-04-06
  13. Divergent Paths to AGI: Symbolic Learning vs. Parametric Scaling and the Shift to Scientific Utilityyoutube · 2026-04-06
  14. Keras with JAX Recommended for Optimal AI Developmenttweet · 2026-04-05
  15. Spurious Correlations in Time Series Visualizationtweet · 2026-04-05
  16. Limits of Curve Fitting in Complex Systemstweet · 2026-04-05
  17. Symbolic Compression and Extreme Generalization in Scientific Discoverytweet · 2026-04-05
  18. Fine-tuning Gemma on TPU v5 with Kinetic, Keras, and JAXtweet · 2026-04-05
  19. JAX: Exemplar of Efficient ML Framework Designtweet · 2026-04-03
  20. Keras Kinetic: Simplified Cloud TPU/GPU Execution for Machine Learningtweet · 2026-04-03
  21. Keras Kinetic and Cloud TPUs for LLM Fine-Tuningtweet · 2026-04-03
  22. Established Companies to Benefit Most from AI Integrationtweet · 2026-04-02
  23. Adobe Podcast Recognized as a Leading AI Producttweet · 2026-04-02
  24. PokeeClaw: Transitioning Local AI Assistants to Enterprise-Grade Productiontweet · 2026-03-30
  25. Humanity’s Chess Mastery Accelerated by Cognitive Infrastructuretweet · 2026-03-29
  26. Rethinking Intelligence as an Optimality-Bound Conversion Ratiotweet · 2026-03-29
  27. Cognitive Agency: The Future Class Divide in an AGI Worldtweet · 2026-03-28
  28. Removing Task-Specific Prompts for General AItweet · 2026-03-27
  29. Beyond Deep Learning: Building Optimal AI with Symbolic Program Synthesisyoutube · 2026-03-27
  30. ARC-AGI Benchmark Roadmap and Design Philosophytweet · 2026-03-26
  31. Redefining AGI: Beyond Benchmarks, Towards Human-Level Learning Efficiencytweet · 2026-03-25
  32. ARC-AGI-3 Benchmarks Agentic AI on Novel Interactive Reasoning Taskstweet · 2026-03-25
  33. ARC-3: A Benchmark for Micro-AGIyoutube · 2025-10-24
  34. Companion Notebooks for Deep Learning with Python (Third Edition) Facilitate Practical Applicationgithub_readme · 2025-09-18
  35. François Chollet: LLMs Are Memorization Engines, Not Intelligence — And AGI Is Still Decades Awayyoutube · 2025-07-23
  36. Chollet's Case Against Scaling: Why Fluid Intelligence Requires Program Search, Not Bigger Modelsyoutube · 2025-06-16
  37. Namex: Streamlining Python Package API Managementgithub_readme · 2025-05-26
  38. ARC-AGI-1: A Grid-Based Benchmark Designed to Test Human-Like Fluid Intelligence in AI Systemsgithub_readme · 2025-04-04
  39. ARC-AGI 2: A New Benchmark for Fluid Intelligence in AIyoutube · 2025-03-24
  40. Challenging Deep Learning for Program Synthesis and Strong Generalizationyoutube · 2025-03-23
  41. Rethinking AI Benchmarking: The Shift to Generalizable, Architecturally-Agnostic Intelligenceyoutube · 2025-01-09
  42. LLMs Are Stuck at System 1: Why Program Synthesis + Deep Learning Is the Path to AGIyoutube · 2024-10-12
  43. Large Language Models: Interpolation Not Intelligenceyoutube · 2024-06-24
  44. Challenging the LLM Hype: The ARC Prize for True AI Generalizationyoutube · 2024-06-11
  45. Active Inference is Key to LLM Performance on Novel Tasksyoutube · 2024-05-03
  46. Keras Ecosystem: Resources for Deep Learning Developmentgithub_readme · 2024-02-12
  47. Keras Blog: Contribution Guidelines and Technical Detailsgithub_readme · 2023-11-01
  48. LLMs as Vector Program Databases for Emergent NLP Capabilitiesblog · 2023-10-09
  49. Deep Learning Requires Interpolative Problems and Abundant Data for Effective Generalizationyoutube · 2021-09-22
  50. Pure Python Nelder-Mead for Environments with Limited Library Supportgithub_readme · 2021-04-24
  51. Deep Learning & Generalization: Interpolation vs. Program Synthesisyoutube · 2021-04-16
  52. Keras Model Repository Deprecation and Usagegithub_readme · 2020-10-01
  53. Rethinking AI: Intelligence as Skill Acquisition, Not Skill Itselfyoutube · 2020-08-31
  54. Challenging the AI "Intelligence Explosion" Narrativeyoutube · 2019-09-14
  55. Seamlessly Interfacing NumPy and Keras Backend Operationsgithub_gist · 2019-04-06
  56. TensorFlow Keras Fails on DeferredTensor Conversion in Hybrid Imperative-Symbolic Seq2Seq Modelsgithub_gist · 2018-10-05
  57. Keras RemoteMonitor Enables Real-Time Metric Visualization via Flask APIgithub_readme · 2018-02-02
  58. Keras RNN API Delivers 15% Faster Stacked LSTM Training on CPUgithub_gist · 2017-09-21
  59. Compact Xception Variant for 200x200 Images with 100 Classes, Sans Residualsgithub_gist · 2017-05-03
  60. Keras Code Snippets for Linear and Logistic Regression with Regularizationgithub_gist · 2016-08-13
  61. Keras Fine-Tuning Script for Small Image Datasets with VGG16 Backbonegithub_gist · 2016-06-06
  62. Keras Transfer Learning Tutorial for Cats-vs-Dogs Using VGG16 Bottleneck Featuresgithub_gist · 2016-06-06
  63. Keras Data Augmentation CNN Achieves High Accuracy on Small Image Datasets via Transfer Learning Principlesgithub_gist · 2016-06-06
  64. Keras Functional API: Layers Callable on Tensors for Graph Modelsgithub_gist · 2016-03-12
  65. Theano Functions Enable Extraction of Intermediate Keras Model Activationsgithub_gist · 2015-05-28
  66. Scientific Progress is Linear, Not Exponentialblog · 2012-08-10
  67. Designing Web Platforms for Enhanced Creativity and Intelligenceblog · 2011-04-30
  68. The "Piano-Playing-Cat Paradigm" and the Wasted Potential of the Internetblog · 2010-12-05