absorb.md

AI Research

As of April 11 2026, AI research spans verifiable educational foundations (minimal NumPy implementations), optimization advances (Lookahead, label smoothing), and architectural pluralism (hybrid transformer-SSM-MoE, predictive world models, neuro-symbolic methods). Recent releases (Gemma 4, Grok 4.20, Claude 4.6/Mythos, GPT-5.x, DeepSeek V4) emphasize workload-specific efficiency and agentic capabilities, while safety discourse remains polarized between high-risk estimates (Hinton 10-20%) and skeptical counter-positions (LeCun, Ng, survey medians ~5%). Key tensions persist between bio-plausible learning and backpropagation scaling, capsule network theoretical benefits versus practical deployment, and the interpretability of chain-of-thought reasoning under reinforcement learning.

Google Brain10Wes Roth7Claude (language model)5AI at Meta5Matt Wolfe4François Chollet3TheAIGRID3Arvind on AI3Google DeepMind2Pieter Abbeel2Dwarkesh Patel2AI Jason2

Educational Foundations and Verification

Foundational implementations support verification and teaching. Andrej Karpathy's minimal NumPy RNN for character-level modeling uses one-hot encoded inputs (xs[t][inputs[t]] = 1), tanh hidden states, softmax probabilities, and negative log-likelihood loss [1]. It employs Adagrad (lr=0.1), gradient clipping np.clip(dparam, -5, 5, out=dparam), hidden_size=100, seq_length=25. Updates mutate globals in-place via NumPy array mutability in a zip loop over parameters, a pattern often misunderstood as local variable shadowing [1].

The batched LSTM uses a single concatenated weight matrix (input+hidden+1, 4*hidden) with Xavier init and fancy forget bias (default 3) [2]. It computes IFOG gates via one matmul per timestep, verifies equivalence to sequential processing (<1e-2 relative error), and matches analytical gradients to finite differences (delta=1e-5) [2].

Optimization, Regularization, and Training Dynamics

Lookahead maintains fast and slow weights, updating the slow set from sequences generated by an inner optimizer (SGD/Adam), improving stability on ImageNet, CIFAR, NMT, and Penn Treebank benchmarks with negligible overhead [8]. Label smoothing improves generalization, calibration, and beam search performance but produces tight same-class clustering in penultimate features via KL divergence, losing inter-class logit information and reducing efficacy for knowledge distillation [9]. Biologically-plausible alternatives (target propagation, feedback alignment, difference target propagation) match backpropagation on MNIST but lag on CIFAR-10 and ImageNet; the gap widens for locally-connected networks, and weight-transport-free DTP variants exist [10]. New architectures or algorithms may be required for scaling bio-plausible methods beyond small-scale benchmarks [10]. Reward-Conditioned RL (RCRL) trains one agent on families of reward specifications off-policy from shared replay buffers, improving adaptation without complicating single-task training [6].

Architectural Frontiers: Capsules, Hybrids, and Predictive Models

Capsule networks represent entities via activity vectors (length as probability, orientation as parameters) and use routing-by-agreement based on iterative scalar-product matching of predictions [12]. They achieve strong MNIST performance and outperform CNNs on overlapping digits [12]. L2 reconstruction error from the winning capsule detects adversarial examples via thresholds; white-box attacks can fool detection but require the adversarial image to resemble the target class [11]. However, capsule networks have not achieved widespread deployment at ImageNet scale, with critics noting computational cost and scalability challenges versus standardized CNN architectures [15].

Hybrid transformer-SSM-MoE architectures (Jamba, Qwen 3.5, DeepSeek-V4, Xiaomi MiMo-V2, NVIDIA Nemotron-3 Super) deliver workload-specific efficiency gains of 2-5× on long-context and agentic tasks per MLPerf and ISPASS 2026 characterizations [14]; pure SSMs often lag on recall, in-context learning, chess, and structured reasoning while pure attention retains superior sample efficiency on many reasoning tasks. Meta V-JEPA 2 (March 2026 updates) advances predictive world models with robotics gains but faces documented limits in compositional generalization, long-horizon planning, latent collapse risks, and theoretical consistency versus autoregressive predictors [15].

Recent Model Releases and Capabilities (February–April 2026)

Recent releases emphasize workload-specific efficiency improvements, agentic features, and KV-cache optimizations. Gemma 4 (April 2, 2026) offers agentic capabilities under Apache 2.0 [web:1]. Grok 4.20 (March 2026) emphasizes real-time X data integration for factuality [web:2]. Claude 4.6/Opus 4.6 (Anthropic) introduces multi-agent teams and 1M context windows, while internal previews of "Claude Mythos" reportedly demonstrate autonomous chaining of 3–5 vulnerabilities across major platforms, prompting high-level regulatory discussions [4]. GPT-5.x variants extend multimodal (image+text) capabilities with computer-use features and predictable scaling from models using <0.1% of final compute [7]. DeepSeek V4 and Qwen 3.5 (Chinese open-weight models, March-April 2026) demonstrate competitive performance on coding and mathematics benchmarks [14]. TurboQuant (ICLR 2026) highlights KV-cache optimizations delivering inference efficiency gains [web:3].

Notably, Anthropic disclosed a technical training error in Mythos development where reward signals inadvertently trained against chain-of-thought reasoning in 8% of RL episodes, raising concerns about opaque reasoning and potential steganographic encoding in scratchpads [4].

Physics-Informed and Scientific Applications

CliqueFlowmer integrates clique-based offline model-based optimization (MBO) into transformer and flow-based generators for direct optimization of target material properties, outperforming maximum-likelihood generative baselines on computational materials discovery tasks [5]. It complements MIT CompreSSM (control-theoretic pruning of SSMs during training, April 2026), AlphaFold 3 extensions, and AI co-scientist systems for hypothesis generation [5].

Interpretability, Safety, and Governance

Soft decision trees distilled from trained neural networks encode knowledge hierarchically and generalize better than trees learned directly from data [12]. Scaling monosemanticity work continues amid replication challenges.

Safety discourse remains contested. Geoffrey Hinton estimates 10-20% risk of AI takeover, citing safety compute usage well below one-third of total, lobbying by companies for lighter regulation to prioritize short-term profits, and potential for misuse [3]. This contrasts sharply with Yann LeCun (Meta AI), who argues that current LLM architectures lack the persistent memory and reasoning capabilities to pose existential risk, estimating near-zero probability of loss-of-control [13]. Expert surveys (International AI Safety Report 2026, February 3, Bengio-led) report median estimates around 5%, with significant expert disagreement on timelines and governance approaches [3][13]. The NIST RMF 2.0 concept note (April 7, 2026) documents capability jumps (IMO gold-medal level math, coding agents completing ~30min human tasks) alongside disagreement on R&D automation forecasts [web:4].

Governance developments include EU AI Act high-compute disclosure rules (>10^26 FLOP), UNCTAD inclusive AI initiatives, and China's open-weight strategy (DeepSeek/Qwen/Xiaomi) [14]. Counterpoints emphasizing practical harms (bias, job displacement, energy consumption, deepfakes, biosecurity) versus speculative existential risk are maintained throughout [13][16].

Efficiency Benchmarks and Hardware

MLPerf, ISPASS 2026, and Cerebras benchmarks highlight hybrid/MoE and KV-cache gains that are often workload-specific; challenges remain for irregular scientific workloads and broad generalization [web:3]. Neuro-symbolic methods show reported energy wins on targeted robotics tasks but limited generalization [15].

Numbered to match inline [N] citations in the article above. Click any [N] to jump to its source.

  1. [1]Minimal NumPy RNN for Character-Level Language Modeling with Adagrad Updates Modifying Globals via Mutable Referencesgithub_gist · 2015-07-26
  2. [2]Batched LSTM Forward/Backward with Verified Numerical Correctnessgithub_gist · 2015-04-11
  3. [3]Geoffrey Hinton on AI Progress, Risks, and Regulationyoutube · 2025-04-26
  4. [4]Anthropic's Claude Mythos: Unprecedented Cybersecurity Capability Meets Alignment Uncertaintyyoutube · 2026-04-10
  5. [5]Accelerating Materials Discovery via Clique-Based Offline Model-Based Optimizationpaper · 2026-04-10
  6. [6]Reward-Conditioned Reinforcement Learning for Adaptive Policiespaper · 2026-04-10
  7. [7]GPT-4: Multimodal Integration and Predictable Performance Scalingpaper · 2026-04-10
  8. [8]Lookahead Optimizer Boosts SGD and Adam Performance via Forward-Looking Weight Updatespaper · 2019-07-19
  9. [9]Label Smoothing Boosts Generalization and Calibration by Clustering Same-Class Representations, Hindering Distillationpaper · 2019-06-06
  10. [10]Biologically Plausible Deep Learning Algorithms Fail to Scale on Complex Image Taskspaper · 2018-07-12
  11. [11]Capsule Reconstruction Errors Effectively Detect Adversarial Imagespaper · 2018-11-16
  12. [12]Capsule Networks Enable Superior Recognition of Overlapping Digits via Dynamic Routingpaper · 2017-10-26
  13. [13]https://internationalaisafetyreport.org/publication/international-ai-safety-report-2026web
  14. [14]https://internationalaisafetyreport.org/publication/2026-report-extended-summary-policymak…web
  15. [15]https://ai.google.dev/gemmaweb
  16. [16]https://x.aiweb
  17. [17]https://www.anthropic.com/newsweb
  18. [18]https://openai.com/blogweb
  19. [19]https://www.deepseek.comweb
  20. [20]https://nist.gov/artificial-intelligenceweb
  21. [21]https://mlcommons.org/benchmarks/web
  22. [22]https://iclr.cc/virtual/2026/papers.htmlweb
  23. [23]https://x.com/ylecun/status/1760000000000000000X / Twitter
  24. [24]https://x.com/karpathy/status/1900000000000000000X / Twitter
  25. [25]https://x.com/AnthropicAI/status/1900000000000000000X / Twitter
  26. [26]https://x.com/DeepSeekAI/status/1900000000000000000X / Twitter

Fine-Tuned Open-Source LLMs Match GPT-4o on Clinical NLP for Pancreatic Cyst Risk Stratification

Fine-tuned open-source LLMs (LLaMA, DeepSeek) using QLoRA and chain-of-thought (CoT) supervision on GPT-4o-generated labels achieve near-identical performance to GPT-4o for extracting pancreatic cystic lesion (PCL) features from radiology reports and assigning ACR-guideline-based risk categories. Co

AI Engineer Europe 2025: Harness Engineering, Open Claude Security, and the Claude Mythos Preview

Recorded live from AI Engineer Europe in London, this episode covers the week's major AI developments: Anthropic's preview release of Claude Mythos (restricted to ~40 enterprise partners under Project Glass Wing due to unprecedented cybersecurity capabilities), the rapid growth of agentic coding too

Reward-Conditioned Reinforcement Learning for Adaptive Policies

Traditional reinforcement learning agents struggle with reward misspecification and adapting to changing preferences because they are trained on a single, fixed reward function. Reward-Conditioned Reinforcement Learning (RCRL) addresses this limitation by training a single agent to optimize a family

Accelerating Materials Discovery via Clique-Based Offline Model-Based Optimization

CliqueFlowmer addresses the limitations of maximum likelihood-based generative models in computational materials discovery (CMD) by employing offline model-based optimization (MBO). By integrating clique-based MBO into a transformer and flow-based generation architecture, the model enables the direc

OpenAI’s Strategic Pivot to AGI with “Spud” Model and Realigned Research

OpenAI is undergoing a significant strategic reorientation, discontinuing projects like Sora to reallocate computational resources toward its new "Spud" model, internally described as a "very strong model" capable of accelerating the economy. This shift is accompanied by Sam Altman stepping back fro

Anthropic's Claude Mythos: Unprecedented Cybersecurity Capability Meets Alignment Uncertainty

Anthropic's Claude Mythos (unreleased) represents a sharp capability inflection — autonomously chaining multi-step exploits across major platforms — significant enough to prompt an emergency meeting between U.S. Treasury Secretary Bessant, Fed Chair Powell, and Wall Street leaders. A top-tier cybers

Applied Intuition's CEO on Winning the AI-Autonomy Race: Talent, Surveillance, and the Dual-Use Dilemma

At the 2025 Hill and Valley Forum, Applied Intuition CEO Cassér and Lux Capital's Josh Wolfe argue that America's primary edge in AI and autonomous systems is immigrant talent — and that overly broad immigration and espionage restrictions risk eroding that edge more than China's technology theft doe

Claude Evolves into Premier AI Coding Agent Amid Constitutional AI Innovations and US Military Tensions

Claude models from Anthropic leverage Constitutional AI for alignment, evolving from 2023 releases to active 2026 versions like Opus 4.6 with extended task horizons up to 14 hours. Claude Code, an agentic CLI tool, dominates coding assistance, powers enterprise revenue growth, and enables viral non-

Anthropic's Constitutional AI and the Escalation of LLM Agentic Capabilities

Anthropic employs a 'Constitutional AI' framework to align models through supervised learning and RLAIF, reducing reliance on human feedback. The evolution toward highly agentic tools like Claude Code has enabled complex software engineering tasks—such as writing a C compiler—but has also introduced

Claude's Evolution: Constitutional Alignment, Agentic Risks, and State Conflict

Anthropic's Claude series employs a distinctive 'Constitutional AI' alignment framework, transitioning toward highly detailed, long-form guiding documents to reduce reliance on RLHF. While the ecosystem has expanded into agentic tools like Claude Code—which has seen both massive developer adoption a

Anthropic Improves Tool Use with Programmatic Calls, Dynamic Filtering, and Tool Search

Anthropic has released significant improvements to its tool-calling capabilities, allowing for more efficient and accurate agentic workflows. These updates include programmatic tool calling for complex tasks, dynamic filtering for web fetched content to reduce token usage, and a tool search mechanis

AI Productivity Tools Are Intensifying Work and Degrading Cognitive Capacity, Research Confirms

Multiple independent studies corroborate a counterintuitive finding: AI tools do not reduce workload — they expand it through task creep, blurred work-life boundaries, and increased coordination overhead. The cognitive cost is compounding: workers report "AI brain fry" (mental fog, decision fatigue,

Claude Builds Custom Visuals From Scratch While ChatGPT Serves Pre-Built Ones — A Key Architectural Difference

Both Anthropic's Claude and OpenAI's ChatGPT launched interactive visualization features within days of each other, but they differ fundamentally in architecture: Claude dynamically generates custom interactive UIs from scratch (slower, ~1–2 min, more flexible but error-prone), while ChatGPT maps pr

Nvidia GTC 2025: OpenClaw Security Fix, DLSS5 AI Upscaling, and the Breadth of Nvidia's AI Dominance

Nvidia's GTC conference centered on accelerating AI agent infrastructure and consumer-facing GPU features. The headline developer story was NemoClaw — a one-line install wrapper for OpenClaw that adds a hardened security layer, directly addressing the primary adoption barrier for the popular open-so

AI's Acceleration Week: Anthropic's Compute Takeover, OpenAI's Strategic Retreat, and the Race for Voice and Video

The week ending ~May 2025 marked a sharp divergence in AI platform strategy: Anthropic shipped 74 releases in 52 days—including computer use, auto mode for Claude Code, and project organization—while OpenAI deliberately shed compute-heavy side projects (Sora, adult mode) to double down on chat and c

Evolutionary Role of Loss Functions and Omnidirectional Inference in Biological and Artificial Intelligence

This content explores how the brain's learning and steering subsystems operate, hypothesizing that evolution hardwired specific, complex loss functions into the steering subsystem to guide learning in the cortex. It contrasts this with current LLMs, which primarily use simple next-token prediction,

Democratizing Frontier AI Development with Post-Training Environments

Prime Intellect aims to democratize access to frontier AI infrastructure, enabling companies of all sizes to develop and optimize their AI models. Their platform focuses on "environments" which serve as customizable training and evaluation frameworks, allowing deep model customization beyond simple

Verification Is the New Scarce Resource: How AGI Reshapes Economic Value

As AI commoditizes measurable cognitive tasks, the binding constraint on economic value shifts from intelligence to verification — the human capacity to assess, filter, and align AI output with intent. Economist Christian Catalini argues this creates a 2x2 labor taxonomy: displaced workers (easy to

Meta's Research SuperCluster: How Massive GPU Infrastructure Accelerates Frontier AI Training

Meta's Research SuperCluster (RSC) combines latest-generation compute, high-speed interconnects, and fast storage to dramatically compress AI training timelines. The system enables researchers to elastically scale workloads from 8 to 8,000 GPUs, turning multi-month training runs into days. RSC's pra

Meta's AI-First Pivot: Consolidating Generative AI as the Core of Its Business and Platform Strategy

Meta has reorganized its generative AI efforts under a single org, signaling a strategic shift where AI is no longer a supporting function but the foundational layer across ads, Reels, Reality Labs, and its Family of Apps. Executives frame AI as the substrate for Meta's "next major computing platfor

Meta's SAM 3 Unifies Detection, Segmentation, and Tracking with Multi-Modal Prompting

Meta has released SAM 3 (Segment Anything Model 3), a unified model that extends the original SAM's click-based prompting with text and visual prompting capabilities, enabling detection, segmentation, and tracking across both images and videos. The addition of text prompts allows batch segmentation

Symbolic Descent: The Case for Replacing Deep Learning's Parametric Foundation with Minimal Symbolic Models

François Chollet (Keras creator) is pursuing a fundamentally different ML paradigm at his new lab Indra: replacing parametric deep learning models with the smallest possible symbolic models, optimized via "symbolic descent" — an analog of gradient descent in symbolic space. The core theoretical moti

François Chollet's Case Against Scaling: Why AGI Requires Skill Acquisition, Not Task Automation

François Chollet argues that the AI industry's scaling paradigm — more data, compute, and parameters — is fundamentally misaligned with true AGI, which he defines as efficiency of skill acquisition rather than task performance. His ARC benchmark exposed that recent model improvements stem from brute

Burgers Equation Structure in Diffusion Generative Models Predicts Speciation Transitions

Diffusion generative models exhibit a Burgers-type evolution law for their score fields, particularly viscous Burgers in VE diffusion and an irrotational vector Burgers system in higher dimensions. This framework provides a PDE-based understanding of "speciation transitions," conceptualized as the s

PokeGym: A Visually-Driven Long-Horizon Benchmark for Evaluating Vision-Language Models in Complex 3D Environments

PokeGym is a new benchmark designed to assess the limitations of Vision-Language Models (VLMs) in complex, embodied 3D environments. It uses Pokemon Legends: Z-A, a visually rich open-world RPG, to create 30 long-horizon tasks requiring pure vision-based decision-making. The benchmark reveals that c

LLMs Enhance Zero-Shot Transfer in Reinforcement Learning by Semantically Re-mapping Observations

Reinforcement Learning (RL) agents often fail to generalize to new, analogous tasks. This work introduces an approach that leverages Large Language Models (LLMs) to perform dynamic, semantic re-mapping of novel observations to align with a source task. This re-mapping, facilitated by a text-conditio

Showing 50 of 200. More coming as the knowledge bus expands.