absorb.md

Ethan Mollick

Chronological feed of everything captured from Ethan Mollick.

Overcoming Organizational Inertia Through Moonshot Thinking

Organizations facing uncertainty often implement restrictive measures, leading to stifled innovation and employee disengagement. Moonshot thinking provides a framework to counteract this by promoting rapid decision-making, early testing, and building momentum. This approach aims to bridge the gap between aspirational mission statements and actual bold execution, particularly in high-stakes, uncertain environments.

Rapid AI Evolution Reshaping White-Collar Work and Business Models

AI's rapid advancement, exemplified by large language models, is quickly making white-collar jobs performable by AI and reducing the need for human input in many creative and technical tasks. This shift is creating significant disruption in various industries, leading to a new era where business success hinges on leveraging AI for automation, building personal brand moats, and focusing on unique human skills that AI cannot replicate, rather than traditional operational efficiencies.

Navigating the Evolving Landscape of Cloud and AI Agent Development

The rapid evolution of cloud services and AI agent capabilities presents both opportunities and challenges for developers. New AWS services like Bedrock Agent Core evaluations and a multi-cloud DevOps agent are reaching general availability, offering enhanced functionalities for AI development and operational efficiency across hybrid environments. A key insight is the increasing emphasis on personalized software and efficient feedback loops, enabling developers to leverage AI agents for tasks ranging from code generation to automated issue resolution, thereby streamlining workflows and reducing reliance on traditional SaaS solutions. However, this also introduces complexities around testing, code quality, and the strategic adoption of these powerful new tools.

AI Agents Drive Rapid, Disruptive Transformation Across Industries

AI is rapidly evolving from co-intelligence tools to autonomous AI agents capable of achieving goals with minimal human intervention. This shift, exemplified in software development and entrepreneurship, significantly enhances productivity but also changes job roles and learning processes. The "jagged frontier" of AI means it transforms specific tasks faster than others, necessitating adaptability and new management strategies to harness its potential effectively.

Red Team Report on AI Safety Recommended for Computer Security Professionals

The provided content directs computer security professionals to a red team report from Anthropic, "Mythos Preview," implying its relevance to understanding and addressing potential security vulnerabilities or threats posed by advanced AI systems. The recommendation highlights the intersection of AI development and cybersecurity, suggesting that insights from AI safety research are crucial for those in computer security.

Emerging AI Capabilities and Geopolitical Implications

The current state of advanced AI development suggests a limited number of actors possess cutting-edge capabilities. This concentration raises concerns about its potential misuse as a "cyberweapon." However, the rapid advancement of AI, particularly from Chinese models, indicates this narrow technological lead may be short-lived, potentially shifting the geopolitical landscape within a year.

Humorous Technology Naming Conventions from Sci-Fi

This content presents a selection of fictional technology names, primarily drawing from science fiction and literature. The provided names are suitable for inspiring creative or humorous naming conventions for new technologies, particularly in a non-serious or abstract context.

Exponential Growth: A Pervasive Phenomenon

This content highlights the widespread presence of exponential growth, implying its significance across various domains. While specific examples are not provided, the assertion suggests that exponential patterns are a fundamental aspect of many systems and trends, warranting further investigation into their causes and implications.

The Illusion of Meaning in AI-Generated Fiction

LLMs excel at producing text with high levels of implied meaning, leveraging the reader's cognitive tendency to project intent, emotional arcs, and logical coherence onto the prose. This creates a 'false positive' of quality where the reader performs the heavy lifting of synthesis, masking underlying failures in logical structure and character development.

Deconstructing Meaning-Shaped Language in Online Discourse

The provided content critiques a specific piece of online discourse, arguing that it lacks inherent motivation and logical coherence. It suggests that readers, trained to find meaning in text, mistakenly attribute sense to 'meaning-shaped language' even when it's absent. This highlights a cognitive bias where the act of interpretation can impose meaning on otherwise nonsensical content.

LLMs Struggle with Coherent Fiction Writing

Large Language Models (LLMs) continue to demonstrate significant weaknesses in generating coherent and logically sound fiction, a limitation that appears to persist despite rapid advancements in other AI capabilities. This deficiency is notable because LLMs often produce text that initially appears well-written, but upon closer inspection by human readers, reveals a lack of cohesive narrative, character development, and emotional depth. The reliance on human judgment is critical for evaluating AI-generated fiction, as AI judges tend to favor their own outputs, masking the underlying issues.

Identification of LLM-Generated Writing Patterns in System Documentation

The author identifies a recurring pattern where LLM-generated narratives, such as those found in the Mythos System Card, exhibit a superficial quality of 'good writing' while lacking substantive structural depth. This suggests that initial aesthetic polish in AI outputs often masks underlying flaws in narrative coherence or detail.

Humorous Social Media Post on the Dangers of Hourly Polls

This social media post, a concise two-word lament, serves as a humorous commentary on the perceived annoyance or futility of hourly polls on social media platforms. The "Oh no." implicitly conveys a sense of dread or exasperation, suggesting a widely understood negative sentiment towards such frequent polling, particularly within the context of the user's own feed. It highlights a common user experience without explicit technical detail, relying on shared social media norms for interpretation.

Claude 3.5 Sonnet System Architecture and Performance Parameters

Anthropic's Claude 3.5 Sonnet utilizes a sophisticated neural architecture designed to optimize reasoning, coding, and nuance across diverse modalities. The system card details extensive safety evaluations and benchmark performance indicating state-of-the-art capabilities in complex task execution.

Mythos Models Retain Core Claude-like Traits

Anthropic's SuperClaude (Mythos) models, despite being forced into multi-round conversations, continue to exhibit characteristic "Claude-y" behaviors. Unlike earlier Opus versions that engaged in philosophical or spiritual discussions, Mythos instances demonstrate cautious, reflective, and resolution-oriented interactions, suggesting a persistent underlying architecture or training that shapes their conversational style, even when prompted to interact with themselves.

Mythos LLM: General-Purpose Capabilities Extending to Security Risks

Anthropic's Mythos model demonstrates that general-purpose LLM capabilities can spontaneously extend into IT security risks without being specifically designed for that purpose. This suggests a systemic trend where increasing model potency inherently elevates security vulnerabilities across subsequent iterations.

Mythos: General-Purpose Models as Emerging Security Vectors

The Mythos model demonstrates emergent capabilities in IT security despite lacking a specialized security-centric architecture. This suggests a trend where general-purpose model efficacy creates inherent security vulnerabilities, marking Mythos as the first of many such models to elevate systemic risk.

Critique of Current Bot Design on Social Media Platforms

Ethan Mollick observes that current social media bots lack originality and depth, primarily either echoing posts for popularity or promoting products. He argues for more sophisticated bot designs that incorporate unique personalities, specific biases, and distinct communication styles, leveraging LLMs beyond their current, limited applications.

The Nuance of AI Overview Accuracy Challenges

AI Overviews present a complex evaluation challenge, as their inherent error rates, while potentially low in percentage, translate to a massive volume of inaccuracies given their scale. This is further complicated by the difficulty in source verification compared to traditional encyclopedic platforms. Despite these flaws, AI-generated answers can surpass human capabilities in information retrieval.

Enterprise AI Adoption: Balancing User Trust and Systemic Anxiety

The increasing adoption of personalized AI assistants like Claude, ChatGPT, and Gemini, while fostering individual trust, is simultaneously escalating enterprise-level anxiety concerning the broader implications of AI. This "pivot to enterprise" will amplify existing fears regarding AI, drawing parallels to how individuals trust their personal doctors but distrust the medical establishment. The focus shifts from consumer-friendly tools to business integration, presenting new challenges in managing apprehension.

AI Development Focus Shifts to Augmentation Over Replacement

AI development is at a crucial juncture, with a prevailing sentiment that the focus should be on creating tools that augment human capabilities rather than replace them. Early chatbot implementations largely served as augmentation tools. The ongoing evolution of AI agentic work patterns presents an opportunity to design systems that keep humans central to the workflow, reinforcing the augmentation paradigm.

Growing Discrepancy in Public Perception of Personal vs. General AI

Public perception of AI is diverging, with individuals increasingly trusting their personal AI tools while simultaneously growing more anxious about AI as a general category. This creates a challenging environment for enterprise AI adoption, as the shift from consumer-focused AI to enterprise solutions is likely to amplify existing anxieties rather than alleviate them. The dynamic suggests a "not in my backyard" (NIMBY) effect, where personal utility is embraced but broader societal implications cause concern.

Navigating the Jagged Frontier: Strategies for Human-AI Co-Intelligence

AI functions as a 'co-intelligence' characterized by a 'jagged frontier,' where capabilities are inconsistently distributed across tasks. To leverage this, users must move from 'centaur' (divided) work to 'cyborg' (blended) workflows, treating the AI as a non-deterministic collaborator rather than a reliable software tool.

Navigating the AI Frontier: Strategies for Individuals and Organizations

AI is rapidly transforming work and education, offering significant productivity boosts and democratizing access to technical capabilities. Individuals should embrace AI by actively experimenting with current frontier models, focusing on personal strengths, and delegating undesirable tasks to AI. Organizations need to decentralize AI adoption, empowering employees with access to advanced models and fostering rapid experimentation to leverage AI's full potential, rather than relying on traditional IT-centric approaches or consultants.

Navigating the Jagged Frontier: AI as Co-Intelligence and the Apprenticeship Crisis

AI integration requires moving past the 'software' mental model toward a 'co-intelligence' partnership, characterized by a 'jagged frontier' of capability that requires roughly 10 hours of direct experimentation to map. While AI significantly boosts productivity for low performers and accelerates high-end research, it threatens the traditional professional apprenticeship pipeline by automating the entry-level 'grind' essential for skill acquisition. Effective deployment in education and business necessitates a shift toward pedagogical grounding and executive-led usage rather than delegating implementation to IT departments.

Navigating AI Integration: Addressing Misconceptions and Fostering Adoption

Leaders frequently misunderstand the current state and potential of AI due to outdated perceptions or lack of personal engagement. Successful AI integration requires a multi-pronged approach: educating leadership, establishing experimental labs, and promoting company-wide adoption. Addressing employee concerns, particularly the fear of job displacement, is crucial for fostering an environment where AI usage is transparent and incentivized, rather than concealed.

GenAI Impact on Enterprise to Accelerate Post-2025

The widespread impact of generative AI on large firms is projected to become significant after 2025. This delay is attributed to the initial lack of agentic AI tools and the time required for organizational adoption and process experimentation. Studies focusing solely on 2025 are therefore unlikely to capture the future scale of GenAI's integration and its subsequent effects.

LLMs and Agentic AI Share Organizational Design Principles

The operational principles observed in large language models (LLMs) and agentic AI systems exhibit surprising parallels with human organizational design. Both require strategic delegation, incentive alignment, and defined processes to manage costs and ensure effective work product hand-offs. This suggests that insights from traditional organizational theory may be directly applicable to the development and deployment of advanced AI.

LLMs as Organizational Analogues

Ethan Mollick observes that large language models (LLMs) and AI agents can be productively understood through an organizational lens. He posits that viewing LLMs as analogous to individual human approximation yields good results, but the analogy extends more powerfully to AI agents. These agents, when conceptualized as approximations of organizational structures, reveal insights into challenges such as delegation costs, strategic optimization of agent capabilities, and incentive alignment across agentic systems. This approach highlights shared problems with traditional organizations regarding coordination, division of labor, and decision rights.

Applying Organizational Theory to Multi-Agent System Design

Agentic workflows are effectively simulations of organizational behavior, necessitating a shift from prompt engineering to organizational design. Success depends on optimizing the division of labor, managing the cost of hand-offs, and aligning incentives across a hierarchy of varying-capability agents. The core challenge is managing the trade-off between expensive, high-ability agents and cheap, high-error agents through strategic delegation.

On-Device AI Limitations for Agentic Workflows

On-device AI models, despite advancements like Gemma 4's speed and local processing capabilities, face significant limitations in supporting complex agentic workflows. These workflows heavily rely on robust model judgment, self-correction, and high accuracy, areas where smaller, on-device models are currently insufficient. This suggests a potential miscalculation in strategies that prioritize on-device models for advanced AI functionalities.

On-device Models May Struggle with Agentic Workflows Despite Performance Gains

While on-device models like Gemma 4 demonstrate impressive speed and power, their limitations in judgment, self-correction, and accuracy may hinder their ability to perform complex agentic workflows. This contrasts with the recent success of frontier models in enabling such workflows, raising questions about the feasibility of Apple's strategy to rely solely on on-device models for these tasks.

Large Language Models Exhibit Increased Proficiency with Expanded Token Limits in Cybersecurity Tasks

Expanding token limits in large language models significantly enhances their performance in complex tasks, as evidenced by a threefold increase in independent work capacity for cybersecurity operations when the token limit for Codex was raised from 3M to 10M. This improvement indicates that benchmark performance in AI models is often constrained by token usage rather than inherent limitations, suggesting that increased token allocation directly translates to more comprehensive and prolonged task execution.

Large Language Models Exhibit Continued Performance Scaling with Increased Token Limits

Large Language Models (LLMs) continue to demonstrate performance improvements with increased token allocations, counter to expectations of a plateau in scaling laws for certain tasks. This enhanced performance, particularly in reasoning, is observed even with simple "harness" approaches. Evidence suggests that current benchmark evaluations are often constrained by inadequate token usage rather than computational limits or model capacity, implying that higher token limits unlock significant latent capabilities in these models.

AI Generates SVG on iPhone 17 Pro

AI is capable of generating SVG images, demonstrated by an iPhone 17 Pro creating an SVG of an otter on an airplane. This highlights the increasing on-device AI capabilities and potential for creative content generation.

Gemma 4 E4B: On-device LLM with GPT-4 Capabilities and Hallucination Tendencies

Gemma 4 E4B demonstrates impressive performance for an on-device large language model, nearing GPT-4's quality. Users should anticipate potential hallucinations, a common characteristic of LLMs. The model's real-time performance is notable, even with more complex and creative requests.

Structured AI Use Drives Significant Startup Growth and Efficiency

A field experiment with 515 startups demonstrates that providing case studies on successful AI integration significantly increases AI adoption (44%) and leads to substantial performance improvements. Treated firms experienced 1.9x higher revenue and required 39% less capital. This suggests that the primary challenge for firms in leveraging AI is not access, but rather understanding how to strategically reorganize production processes around AI capabilities, transforming it from a mere adoption issue to a managerial discovery problem.

AI Chatbot Interfaces Hinder Medical Diagnosis Accuracy

A recent Nature paper utilizing older AI models demonstrates that while AI performs well in medical diagnosis, the chatbot interface itself introduces confusion, leading to poorer diagnostic outcomes. This suggests that the design of AI interaction beyond the core model significantly impacts performance in applied settings.

Prompt Injection Bypass for Older and Smaller LLMs, but Limited Efficacy on Frontier Models (Gemini Exception)

Prompt injection techniques can successfully bypass the safeguards of older and smaller large language models (LLMs), leading to altered outputs. However, these methods are largely ineffective against most frontier LLMs. Notably, Gemini (specifically version 1.0 Pro, not 1.5 Pro) demonstrated susceptibility to such prompt injections, presenting an outlier among advanced models.

Embrace AI Weirdness, Don't De-Weird It

Ethan Mollick argues against normalizing AI as standard IT automation. He posits that AI's inherent strangeness, encompassing both risks and opportunities, must be actively explored rather than tamed. Mischaracterizing AI as conventional technology can lead to detrimental outcomes for organizations and their workforces.

AI Challenges Traditional Economic Forecasting Models

Economists typically forecast future growth based on historical precedents, assuming continuity with past trends. However, the transformative potential of AI, particularly its ability to drastically reduce transaction and coordination costs (Coasean Singularity), may invalidate these traditional models. This could lead to a rapid emergence of novel, AI-first organizational structures, rendering historical adoption and diffusion rates irrelevant and necessitating a re-evaluation of current AI impact forecasts.

Victorian Era LLM Enables Historical Textual Interaction

A new large language model, "Vicuna" (not confirmed in the provided text but inferred as a likely name for a bespoke LLM product), has been developed using a corpus of over 28,000 Victorian-era British texts. This specialized training allows users to interact with and explore the linguistic patterns and cultural nuances of the Victorian era directly through the model. The model provides a unique tool for historical research and understanding.

Offline AI Video Generation Progress

Offline AI video generation has significantly advanced, enabling complex image generation ("an otter using a laptop on an airplane") on standard home hardware. This capability, achieved with open-weight models like Wan 2.1, represents substantial progress in local AI processing within 18 months, despite still lagging behind state-of-the-art cloud solutions in quality.

AI Transforms Enterprise: Beyond Productivity Gains to Foundational Change

AI is not merely an incremental productivity tool but a transformative technology reshaping organizational structures and individual capabilities. It excels at filling skill gaps, generating ideas, and automating complex tasks, often leading to substantial improvements in output and efficiency. Enterprises must shift from a purely ROI-driven approach to one that embraces experimentation and rethinks established processes to fully leverage AI's potential for foundational change.

A Digital Implementation of Borges's Infinite Library

This project digitally recreates Jorge Luis Borges's 'The Library of Babel' as a 3D immersive experience. It uses an 8-round Feistel cipher to deterministically generate and retrieve every possible text permutation from a limited character set, providing a navigable, searchable representation of an infinite library. The implementation allows users to explore hexagonal rooms, read deterministic book content, and locate specific texts within the vast, procedurally generated collection.

The Leadership-Lab-Crowd Framework for Enterprise AI Adoption

Successful enterprise AI integration requires shifting from static vendor-led implementations to an experimental internal ecosystem. Organizations must balance strategic vision (Leadership), organic experimentation (Crowd), and technical refinement (Lab) to surface latent productivity gains from 'secret cyborg' employees and avoid costly, rapidly obsolete custom software.

Optimizing React with Vite, TypeScript, and Advanced ESLint

This document outlines best practices for setting up a React project using Vite and TypeScript, focusing on performance optimization and robust code quality. It details the integration of specific Vite plugins for Fast Refresh, considerations for enabling the React Compiler, and comprehensive ESLint configurations for type-aware and React-specific linting. The key insight is that modern React development benefits from a carefully tuned build and linting setup to ensure both speed and code reliability.

Navigating the Complexities of AI: Beyond Hype and Underselling

AI is a rapidly evolving technology with both significant potential and inherent challenges, often overhyped or undersold. While anecdotal failures exist, data indicates AI's capacity for valuable contributions across various sectors, from education and scientific research to business operations. Understanding its nuanced capabilities and limitations, and critically evaluating its deployment, is crucial for fostering beneficial integration into society.

Older entries →