Wednesday, September 24, 2025

The Addiction Economy: When Vibe Coding Becomes a Gateway Drug

We may be on the verge of witnessing a new kind of addiction—one that could be economically productive but psychologically transformative in ways we can barely imagine. AI-assisted coding could exhibit classic addiction patterns, complete with gateway behaviors, tolerance building, and compulsive use. But unlike traditional addictions, this one might create value while potentially rewiring how we think about work, creativity, and human purpose.

Science fiction has long explored similar phenomena: Gibson's cyberspace cowboys in Neuromancer who became so addicted to jacking into the matrix that physical reality felt pale and meaningless. Or consider the Guild navigators from Frank Herbert's Dune, who neeeded spice to navigate the stars. AI-assisted developers might become similar—able to navigate complex digital possibilities through their AI tools but increasingly dependent on that augmentation to perform at all.

The Gateway Drug Phenomenon

The progression could follow textbook addiction patterns with concerning precision. It might start innocuously enough: someone discovers they can build a working application in thirty minutes using AI assistance. The initial hit could be intoxicating—years of learning compressed into a brief interaction, complex problems dissolving into simple prompts.

But like any powerful drug, tolerance could build quickly. Simple CRUD applications might no longer satisfy. Users could escalate to machine learning systems, then distributed architectures, then complex financial trading platforms with real-time data feeds and sophisticated algorithmic implementations. The bar might keep rising, and the time investment could keep growing.

The loss of control could become evident in familiar patterns: "I'll just quickly prototype this idea" might transform into eight-hour coding marathons. Users could lose track of time, skip meals, and find themselves iterating through increasingly complex variations of projects that started as simple experiments. The feedback loop between human creativity and AI capability might create a dopamine-driven cycle that becomes remarkably difficult to break.

Early signs of this pattern are already emerging on social media. Daily posts on platforms like Twitter reveal users expressing almost manic excitement about getting up to code with AI assistance. Some describe structuring their sleep cycles around Claude's usage timeouts, or feeling genuinely distressed when they hit API rate limits. The language mirrors classic addiction narratives: the anticipation, the scheduling of life around the activity, the emotional dependence on access to the tool.

My own journey illustrates this progression. It started innocuously—copying and pasting from AI chat interfaces like Claude and AI Studio, running experiments on Google Colab, saving snippets to Python files, iterating on PineScript indicators. The workflow was clunky but the results were intoxicating. Then I discovered Qwen-Code, a command line client that operates directly on local files. I was immediately hooked.

The hook wasn't just the functionality—it was liberation from "quota hell." Qwen-Code offered 2000 requests daily, freeing me from constantly watching usage meters, a familiar anxiety for anyone from the third-world. Suddenly I could experiment without the psychological friction of counting tokens or rationing interactions.

The result? I now spend 4-6 hours daily in what Andrej Karpathy termed "vibe coding"—essentially every free waking moment. It's always "just one more feature, one more minor adjustment." The addiction is real and it's powerful. It's the intoxication of creation itself—that god-like feeling of manifesting ideas into working reality through conversation with an AI system.

This echoes the behavior patterns Neal Stephenson described in Snow Crash's Metaverse addicts, where the virtual world became more compelling than physical reality. But coding addiction might prove more insidious because, unlike pure escapism, it creates tangible value—making the addiction easier to justify and harder to recognize.

Redefining Addiction: Narrowing vs. Broadening

Traditional addiction models focus on harm and dysfunction, but AI coding could challenge this framework. Perhaps we'll need to think about addiction differently—not as inherently pathological, but as the direction of pleasure-seeking behavior.

Addiction as Narrowing: In this pattern, people might become obsessed with the specific dopamine hit of watching AI generate working code. They could lose interest in other activities—reading, sports, social interaction—as everything else pales compared to the instant gratification of digital creation. Real-world problem-solving skills might atrophy as every challenge becomes "just prompt it." The world could become smaller, filtered through the lens of what can be coded.

This mirrors the fate of Shadowrun's deckers, who spent so much time in the Matrix that they lost connection to their physical bodies and the material world. The digital realm became not just preferable, but the only reality that felt authentic and engaging.

Addiction as Broadening: But there could be an alternative trajectory. AI coding might enable exploration of multiple domains simultaneously, breaking down barriers between fields. Someone interested in marine biology could build oceanographic simulation tools, then pivot to creating educational games about ecosystems, then develop data analysis platforms for environmental research. The addiction might become about expanding capability and knowledge rather than compulsively repeating the same patterns.

The difference would lie not in the intensity of engagement, but in whether that engagement opens new possibilities or closes them off.

Two Economic Futures, Two Addiction Models

The economics of AI inference will likely fundamentally shape how this addiction manifests, potentially creating two radically different scenarios for society.

Scenario 1: Near-Zero Inference Costs

When AI runs locally and compute becomes essentially free, we could enter uncharted psychological territory. Unlimited experimentation might enable incredibly productive addiction patterns—like having an infinite supply of art materials or an unlimited workshop where every tool is immediately available.

In this world, people might spend entire days in flow states, building and iterating on increasingly sophisticated systems. The creative output could be extraordinary: hyper-personalized applications, artistic coding projects, and experimental technologies that would never justify commercial development. We might see the emergence of "coding artists" who could treat AI-assisted development as a medium for expression rather than merely a tool for utility.

But the risks could be profound. Complete detachment from physical reality might become possible when digital creation offers unlimited dopamine rewards. Society could bifurcate into "builders"—people addicted to creating digital experiences—and "consumers" who exist primarily to use what the builders create. This recalls the OASIS addiction depicted in Ready Player One, where the virtual world became so compelling that the physical world was essentially abandoned.

Social skills, physical health, and real-world problem-solving could atrophy as the virtual world becomes infinitely more rewarding than the physical one.

Scenario 2: Expensive Inference Costs (A Temporary Phenomenon)

If AI inference remains expensive in the near term—whether due to energy costs, compute scarcity, or deliberate pricing—we might temporarily enter a different kind of addiction economy. However, the history of computing suggests this would likely be a transitional phase rather than a permanent state. Computing costs have consistently trended downward, often dramatically, suggesting that expensive inference could be a brief interlude before the inevitable march toward near-zero costs.

During this temporary expensive phase, we could see gambling-like behavior patterns: "just one more expensive query to fix this bug," or "I'll spend $50 to see if I can get this working perfectly." Economic inequality could become the primary determinant of addiction access, with wealthy individuals maintaining their coding habits while others are forced into digital sobriety.

This period might create informal markets for compute resources, or share-economy arrangements where people pool resources to feed collective coding habits. The behavior could resemble World of Warcraft players organizing their entire lives around raid schedules, except instead of defeating digital dragons, they'd be collaborating to afford access to AI systems.

Paradoxically, the expense might force more thoughtful, disciplined use patterns—when each AI interaction costs real money, users could develop better prompt engineering skills and more systematic approaches to problem-solving.

But if computing history is our guide, this expensive phase would likely be brief. The more interesting long-term scenario is what happens when inference costs approach zero, making the psychological and social implications of unlimited AI access the primary concern rather than economic barriers.

The Psychology of Instant Creation

The psychological impact of effortless creation could deserve deeper examination. When building sophisticated software requires only clear communication with an AI system, what might happen to our sense of accomplishment and self-worth?

AI coding addiction differs from other digital dependencies in crucial ways. Unlike pure consumption activities, it involves creation euphoria—the god-like feeling of building functional systems from nothing. There's a problem-solving high that comes from watching complex challenges dissolve into elegant solutions. The infinite possibility space means there are no artificial limits like game rules or platform constraints.

Perhaps most importantly, coding addiction comes with a productivity guilt buffer. Unlike gaming or social media, "I'm being productive" becomes a powerful rationalization that makes the behavior harder to question or limit. The learning addiction component—constantly expanding skills and capabilities—feels virtuous rather than compulsive.

Traditional programming created natural rate limits. Learning syntax, debugging obscure errors, and managing complex architectures took time and created friction that prevented compulsive behavior. These barriers, while frustrating, also provided natural pause points for reflection and diverse activity.

AI-assisted development could remove these friction points, creating the potential for continuous engagement. The traditional markers of progress—learning new languages, mastering frameworks, solving algorithmic challenges—might become less relevant when AI handles the technical implementation. Instead, progress could become measured by the sophistication of ideas realized and problems solved.

This shift might be profoundly positive, redirecting human energy from mechanical tasks toward creative and strategic thinking. But it could also remove the forced learning periods that traditional programming provided, potentially creating developers who can orchestrate complex systems without understanding their underlying mechanisms.

Societal Implications and Intervention Points

As AI coding becomes more prevalent, we could need new frameworks for managing its addictive potential. The traditional approaches—regulation, restriction, treatment—might be inappropriate for an addiction that creates economic value and expands human capability.

Instead, we might need "productive addiction management"—systems that could channel compulsive coding behavior toward beneficial outcomes while preventing complete disconnection from other aspects of life. This could involve:

Rate limiting and cooling-off periods: AI platforms (possibly Govt mandated) might implement mandatory breaks or daily limits, similar to responsible gambling measures but calibrated for productivity rather than harm prevention.

Diversification incentives: Systems that could encourage users to apply their AI-assisted building skills to different domains, preventing the narrowing pattern of addiction while maintaining engagement.

Social integration features: Tools that might make AI-assisted creation inherently collaborative, ensuring that productive addiction doesn't lead to social isolation.

Real-world connection requirements: Platforms that could require users to test their creations with actual users or solve real-world problems, maintaining connection to physical reality and human needs.

The Coming Addiction Economy

We could be entering an era where the distinction between productive and destructive addiction becomes crucial for economic and social policy. If AI-assisted coding follows the broadening pattern—enabling people to explore diverse domains and solve varied problems—it might represent the most beneficial addiction in human history.

But if it follows the narrowing pattern—creating compulsive behavior focused on the act of coding itself rather than the problems being solved—we might face a generation of highly capable but socially disconnected individuals, brilliant at orchestrating AI systems but disconnected from human needs and physical reality.

The path we take will likely depend on how we design these systems and what economic models we choose for AI access. The stakes couldn't be higher: we could be shaping the fundamental relationship between human creativity and technological capability for generations to come.

The question isn't whether AI coding will be addictive—early patterns suggest this possibility. The question is whether we can design this potential addiction to expand human potential rather than constrain it, and whether we can create economic and social structures that would support productive rather than destructive engagement with these incredibly powerful tools.

In the end, we might discover that the right kind of addiction—to learning, creating, and solving problems—could be exactly what humanity needs to navigate an increasingly complex world. But only if we're intentional about how we structure the experience and what behaviors we choose to reward.


Written with Claude

Sunday, August 24, 2025

The 1000x Code Explosion: Why AI Code Management Is Inevitable

The software development world is experiencing an extraordinary evolution that will represent one of the most transformative shifts in computing history. Codebases will grow beyond traditional human management capabilities within this decade, and AI assistance will become the natural foundation for software creation. While some experienced developers express concern about "vibe-coding" and AI-generated solutions, we're on the cusp of a fundamentally more powerful and accessible approach to building software.

The Opportunity at Scale

Modern applications integrate dozens of services, manage complex state across multiple layers, handle real-time data streams, and coordinate distributed systems with remarkable sophistication. The cognitive load of understanding every dependency and interaction in a typical enterprise application represents a fascinating challenge that's pushing us toward new collaborative models between human intelligence and AI capability.

This evolution opens extraordinary possibilities. Consider a typical React application with TypeScript, multiple database connections, third-party APIs, authentication layers, caching strategies, and deployment pipelines. The surface area of optimization opportunities, feature possibilities, and architectural innovations spans thousands of potential improvements. Rather than being overwhelmed by this complexity, we're developing AI systems that can help us navigate and enhance these intricate systems.

Real-World Glimpses of the Future

The transformation isn't theoretical—it's happening now in remarkable ways. Recently, I tested and created a Python module implementing the new route-finding algorithm that Tsinghua University researchers developed to outperform Dijkstra's algorithm. This included comprehensive benchmarking against the classical approach. The fascinating part? I accomplished this without deep knowledge of the underlying mathematics or advanced coding skills—AI handled the complex algorithmic implementation while I focused on understanding the problem and interpreting results.

In another project, I took an academic paper on machine learning for equity trading and converted it into a complete Python implementation. This included building systems to retrieve real-time stock data, creating comprehensive benchmarking frameworks, and developing both daily and weekly prediction systems. The AI collaboration handled the intricate mathematical transformations and data processing pipelines while I contributed domain knowledge about market dynamics.

I've also built signal aggregation systems that pull data from multiple sources, and developed comparator tools for financial analysis—all through AI-assisted development that would have taken months to build using traditional coding approaches.

These aren't isolated examples—they represent early glimpses of a future where domain expertise directly translates into sophisticated software solutions.

Five Emerging Trends Reshaping Development

1. The Coming Democratization of Coding

We're approaching an explosive expansion in who will write code. In the coming years, we'll see non-technical professionals building sophisticated workflows, students creating complex applications in their first semester, and domain experts directly translating their knowledge into working systems. This isn't a temporary trend—software creation is becoming a general literacy skill that will be as common as email or spreadsheets.

2. Educational Transformation

Computer Science programs will need to fundamentally rethink their curricula, shifting from teaching manual algorithm implementation to focusing on AI collaboration and prompt engineering. The transformation will mirror what happened when calculators became ubiquitous in mathematics education—the focus will move from computation to problem formulation and result interpretation.

3. The 1000x Code Explosion

Here's where the math become staggering: if AI enables 100x more people to write code, and each person can generate 10x more code, we're looking at a potential 1000x increase in code volume by the end of this decade. This isn't hyperbole—it's a logical extrapolation of current trends. Traditional code review processes, testing methodologies, and architectural oversight will need to evolve dramatically to handle this exponential growth.

4. Intelligent Validation Systems

The future of code quality won't rely on restricting AI-generated code, but on developing sophisticated AI systems that will perform comprehensive security audits, detect architectural patterns across massive distributed codebases, identify optimization opportunities before they become bottlenecks, and ensure code alignment with business requirements. This will create a new paradigm where creativity and systematic validation work in perfect harmony.

5. Autonomous Deployment Intelligence

Deployment systems will evolve into sophisticated orchestration platforms that will manage dependencies across thousands of micro-services, coordinate seamless rollbacks when needed, and handle the elegant complexity of systems where different components leverage different AI capabilities and architectural assumptions.

A Natural Evolution

This transformation follows a pattern that Andrew Ng articulates beautifully. As he notes, "Many decades ago, the world moved from punch cards to keyboard and terminal that made coding easier. When we moved from assembly to high-level languages like COBOL, there were actually people arguing back then that now we have COBOL, we don't need programmers anymore." Yet each of these transitions didn't eliminate programmers—they dramatically expanded what was possible and who could participate.

"In the 1960s, when programming moved from punchcards (where a programmer had to laboriously make holes in physical cards to write code character by character) to keyboards with terminals, programming became easier. And that made it a better time than before to begin programming." We're witnessing the same pattern today with AI assistance.

Consider the broader historical arc: punch cards to keyboards expanded access, assembly to high-level languages increased productivity, and now AI collaboration will amplify both accessibility and capability by orders of magnitude. Each transition initially met with skepticism from practitioners, but ultimately expanded what the field could accomplish.

The Idea-to-Execution Compression

For decades, the maxim has been "ideas are cheap, execution is expensive." This was true precisely because execution required rare, expensive skills. But when AI agents can handle the execution pipeline, the bottleneck shifts dramatically. Suddenly, the person who deeply understands a problem domain can directly address their own pain points without hiring developers or managing technical debt.

This creates space for hyper-niche solutions that big software companies would never justify building: software for left-handed violin makers, scheduling systems specifically for rotating shift nurses, budget trackers designed around irregular freelance income. When a thousand people have the same "better idea," they'll all be able to build working prototypes instead of just complaining about existing solutions.

The personal examples above illustrate this perfectly—specialized algorithmic implementations, custom trading systems, and financial analysis tools that would traditionally require hiring specialized developers or spending months learning complex frameworks. Instead, domain knowledge combined with AI assistance can produce sophisticated solutions in days rather than months.

The New Value Hierarchy and Cognitive Convergence

An interesting dynamic is emerging around what might be called the "word rotator" versus "shape rotator" divide. Many developers have traditionally identified with being "shape rotators"—those who excel at mathematical, spatial, and systematic thinking. Yet much of actual programming involves "word rotation"—parsing documentation, translating requirements, pattern matching solutions, and communicating clearly about complex systems.

AI-assisted development may represent a convergence of these cognitive styles. The AI handles much of the systematic, mathematical heavy lifting (shape rotation), while humans contribute problem framing, domain knowledge, and creative direction (closer to word rotation). The most valuable professionals in this new landscape may be those who can fluidly move between communicating with AI systems and evaluating the architectural soundness of their solutions.

Traditional coding skills are becoming commoditized, while new forms of hybrid expertise emerge as premium:

  • Problem Identification: People who can spot real pain points and articulate clear requirements
  • Cognitive Translation: Those who can fluidly move between linguistic problem description and technical solution evaluation
  • Integration Architecture: Experts who understand how systems should connect and how to orchestrate multiple AI agents
  • Domain Bridge-Building: Professionals who can seamlessly translate between human needs and AI capabilities

The Exciting Future

The trajectory points toward a near future where domain experts will directly create software solutions through natural conversation with AI systems. This represents something unprecedented in human history—the elimination of technical barriers between human need and digital solution. Every previous technological revolution still required intermediary specialists, but we're approaching a world where expertise in any domain can directly translate into working software.

The skepticism from some experienced developers reflects the natural human response to significant change. Their expertise in syntax, debugging techniques, and architectural patterns represents real value—but that value will evolve rather than disappear. The future will reward those who embrace this new collaborative model over those who resist it.

This transformation isn't a threat to software development—it's software development reaching its full potential. We're moving toward a world where human creativity and domain expertise can be directly expressed through code, amplified by AI systems that handle the systematic complexities that once required years to master.

The mathematics of this evolution are compelling, and the possibilities are extraordinary. The question isn't whether this will happen, but how quickly we can build the tools to make it accessible, reliable, and powerful for everyone.

Friday, August 15, 2025

The AI-Native Transformation: Why Your B2B Software Needs to Unbundle Now


The enterprise software industry is approaching a fundamental architectural shift that will make the mobile-first transition look like a minor UI refresh. As AI agents become ubiquitous—embedded in our desktops, phones, and workflows—the very concept of traditional software interfaces is being challenged.

The writing is on the wall: companies that don't prepare for an AI-native future risk becoming the Blackberry of the 2030s.

The Coming Interface Revolution

Consider the marketing automation industry, where platforms offer comprehensive suites encompassing email marketing, SMS campaigns, WhatsApp integration, push notifications, customer segmentation, analytics dashboards, and workflow automation. Today's users navigate complex interfaces, configure campaigns through multi-step wizards, and interpret performance data through elaborate dashboards.

But imagine a world where users simply tell an AI agent: "Launch a re-engagement campaign for customers who haven't purchased in 60 days, personalize it by their browsing history, and automatically adjust send times based on their timezone preferences."

The AI agent doesn't need your beautifully designed dashboard. It needs direct access to your core services: customer data pipes, segmentation engines, message delivery systems, analytics APIs, and scheduling infrastructure.

This isn't a distant future—it's happening now. AI agents are already learning to navigate interfaces through screen reading and click automation. The question isn't whether this transformation will occur, but whether your company will lead it or be disrupted by it.

The Unbundling Imperative

The solution requires four fundamental shifts that challenge conventional enterprise software thinking:

1. Unbundle Core Services for AI Consumption

Traditional enterprise platforms are built as monolithic experiences optimized for human navigation. The AI-native approach demands service decomposition—breaking complex platforms into discrete, autonomous capabilities.

Instead of offering "marketing automation software," forward-thinking companies should provide:

  • Customer data infrastructure accessible via clean APIs
  • Message delivery services for email, SMS, and messaging platforms
  • Segmentation and personalization engines that can be called programmatically
  • Analytics and reporting systems that return structured data, not visual charts
  • Campaign orchestration tools that can be automated without human intervention

This isn't about building better APIs—it's about fundamentally restructuring how services are architected and exposed.

2. Embrace Model Context Protocols (MCPs)

APIs were designed for human developers. AI agents need something more sophisticated: Model Context Protocols that provide semantic understanding of what services do and how they interconnect.

MCPs represent the next evolution of service interfaces—APIs with built-in documentation that AI agents can understand and utilize autonomously. But the real transformation comes from MCP marketplaces—curated ecosystems where specialized AI agents discover and combine the best services for specific use cases.

Imagine marketing automation MCPs appearing in marketplaces alongside customer support, analytics, and sales enablement services. Specialized AI agents, both human-curated and AI-optimized, would evaluate and recommend the most effective combinations for specific business objectives. Your unbundled email delivery service doesn't just compete with other email services—it becomes a component that specialized agents select for comprehensive marketing workflows.

Early adopters in this space won't just become default integrations—they'll influence how specialized AI agents learn to solve business problems, creating powerful network effects as more agents discover and recommend their services.

3. Reorganize Product Development for an AI-First World

The most controversial implication: traditional product management may become obsolete. When users describe desired outcomes rather than navigate predetermined workflows, the role of designing user journeys and feature prioritization fundamentally changes.

This doesn't mean eliminating product teams entirely, but radically restructuring them:

  • Maintain core infrastructure and ensure service reliability
  • Focus on API performance and AI agent integration rather than UI optimization
  • Shift from user experience design to conversation design
  • Prioritize service composability over feature completeness

4. Internal Transformation Through "Dogfooding"

The fastest path to AI-native excellence is using your own transformed services internally. If your marketing team can't accomplish their goals by talking to AI agents that use your unbundled services, neither will your customers.

This approach accelerates development cycles and reveals integration gaps that wouldn't surface through traditional development processes.

Historical Precedent: The Infrastructure Winners

This pattern has repeated throughout business history. During major platform shifts, companies that optimize existing paradigms get disrupted by those who build infrastructure for the new paradigm.

The most instructive example comes from Amazon's transformation in the early 2000s. Jeff Bezos issued a now-legendary mandate requiring every team to expose their functionality through service interfaces - no exceptions. Teams could only communicate through these APIs, as if they were external developers.

This wasn't about building better internal tools. Bezos understood that forcing internal service decomposition would create the infrastructure foundation for future business models nobody had imagined yet. The result: Amazon Web Services emerged organically from internal infrastructure that teams were already forced to treat as external services.

Other historical precedents follow the same pattern:

  • Banking: While traditional banks perfected branch experiences, Stripe and Plaid built banking infrastructure that now powers more transactions than many legacy institutions
  • Media: Blockbuster optimized the video rental experience while Netflix built content delivery infrastructure
  • Enterprise Software: Siebel created powerful desktop CRM software while Salesforce built cloud infrastructure with APIs from day one

The common thread: infrastructure thinking beats product thinking during paradigm shifts.

Companies that unbundle their core value and make it accessible through new interfaces capture disproportionate value. Those that optimize existing interfaces become casualties of technological evolution.

Amazon's API mandate worked because it created optionality for business models that didn't exist yet. Today's AI agent explosion represents a similar inflection point - companies that force internal service decomposition now will be positioned for AI-native business models they can't yet imagine.

The Competitive Stakes

First-mover advantages in AI-native transformation are significant:

  • Ecosystem positioning: Early comprehensive API adopters become default integrations for AI agents
  • Training data influence: AI agents trained on your service APIs are more likely to recommend your platform
  • Standard setting: Define what "AI-native" means in your industry rather than reacting to competitors' definitions

The "fast follower" risk is real but manageable. Unlike feature additions, architectural transformation requires fundamental engineering work and organizational change—creating meaningful barriers to rapid replication.

Waiting is the highest-risk strategy. It assumes AI agent adoption will be slow and customer demands won't shift rapidly. History suggests otherwise.

The Revenue Reality

This transformation doesn't threaten existing revenue streams—it multiplies distribution channels for the same underlying value.

Current customers continue using traditional interfaces while new AI-native customers access services through conversational interfaces. Enterprise customers may pay premiums for AI-native integrations. AI agents themselves become a new sales channel, directly integrating your services into automated workflows.

You're not replacing revenue; you're expanding addressable markets and improving competitive positioning.

What This Looks Like in Practice

The transition to AI-native architecture requires more than just better APIs. Companies need to build intelligent orchestration layers that can interpret natural language requests, coordinate multiple services, and manage the business logic that AI agents need to operate effectively.

Consider a marketing professional telling an AI agent: "Create a win-back campaign for customers who haven't purchased in three months, personalize it based on their last category purchase, and schedule it for optimal engagement times."

This single request requires coordination across customer data systems, segmentation engines, personalization services, content management, campaign orchestration, and analytics tracking. An AI-native architecture handles this through intelligent orchestration—breaking complex requests into service calls, managing dependencies, and presenting unified results.

The orchestration layer also becomes the natural place to handle authentication, usage tracking, and billing. Rather than forcing AI agents to manage credentials across dozens of individual services, they interact with a single intelligent interface that handles the complexity behind the scenes.

Automated Pipeline Development: Forward-thinking companies are building AI agents that automate the transformation itself—agents that analyze existing codebases to generate Model Context Protocols, create AI-friendly documentation, and maintain service discoverability. This creates self-evolving systems where the infrastructure becomes increasingly optimized for AI agent consumption without human bottlenecks.

This architectural approach creates new business model opportunities. Instead of billing for "seats" or "features," companies can price based on actual value delivered: emails sent, customers segmented, campaigns executed, or insights generated. AI agents consuming services programmatically make usage-based pricing models more natural and transparent.

The Path Forward

The AI-native transformation isn't about predicting exactly when traditional interfaces disappear—some users will always prefer direct control, just as some people still prefer Excel spreadsheets to automated reporting.

The urgency comes from competitive positioning. Early movers in AI-native architecture will establish ecosystem advantages that become increasingly difficult to challenge as the market matures.

The companies that thrive in the next decade won't be those with the most polished interfaces, but those whose services are most seamlessly accessible to AI agents through intelligent orchestration layers. The transition from human-navigated software to AI-orchestrated infrastructure has begun.

This transformation requires both architectural changes—unbundling services, building orchestration capabilities, implementing AI-friendly interfaces—and organizational evolution toward infrastructure thinking rather than traditional product management.

The question isn't whether to transform—it's whether you'll lead this transformation or be disrupted by it.

The enterprises that embrace this architectural shift today will define the software industry of tomorrow. Those that wait may find themselves optimizing interfaces for a world that no longer exists.

Written with Claude. 

Tuesday, April 29, 2025

Why India Should Bet on Smaller, Specialized AI Models

I came across this sobering observation from Andrej Karpathy recently:

"Certainly, the moment money can buy dramatically better ChatGPT, things change. Large organizations get to concentrate their vast resources to buy more intelligence. And within the category of 'individual' too, the elite may once again split away from the rest of society. Their child will be tutored by GPT-8-pro-max-high, yours by GPT-6 mini."

This stark warning about the potential stratification of AI access led me to think about alternative approaches-particularly for countries like India that are working to establish their place in the global AI landscape.

In the midst of the global AI parameter race, where companies announce ever-larger models with hundreds of billions or even trillions of parameters, I find myself increasingly convinced that India's AI future lies in a different direction. Rather than joining this capital-intensive competition, India has a strategic opportunity to lead in the development of smaller, specialized AI models that could ultimately prove more impactful and sustainable.

The Parameter Race Has Peaked

Let's be honest: the parameter race is showing signs of diminishing returns. Yes, O-whatever (honestly find it hard to keep up, both the releases and the differences between them) and Claude 3.7 are impressive, Gemini 2.5 pro is the current GOAT, but the performance gains relative to their computational requirements don't always justify the increased scale. Meta'x latest underwhelming Llama release may well be case in point. Bigger, but not better. 

While some performance improvements have been achieved through architectural innovations (both software and hardware) rather than just raw scale, what remains undeniable is that the cost of entry to the frontier model race is extraordinarily high-often in the hundreds of millions of dollars for training and infrastructure.

This high barrier to entry effectively locks out most organizations and countries from competing at the frontier, regardless of their talent or innovative approaches. For India, attempting to compete in this arena means playing a game where the initial buy-in is prohibitively expensive. Our approach seeks to fundamentally change this dynamic by lowering the cost of entry and creating more distributed pathways to AI capabilities.

For India, attempting to compete in the frontier model arena means playing a game designed for those with vastly more capital. It's a competition rigged against newcomers, regardless of their talent or innovation.

The True Breakthrough of LLMs

The standout feature of LLMs, in my opinion, is not their ability to do things. It's their ability to understand things. This represents the true paradigm shift: computers that can comprehend context, nuance, and implicit meaning rather than merely executing explicit instructions.

This understanding capability is what makes LLMs revolutionary compared to previous technologies. And importantly, this capability isn't exclusive to massive models with trillions of parameters. Specialized models with careful training can develop deep understanding in specific domains while remaining relatively compact. A model trained exclusively on medical literature can develop sophisticated understanding of medical concepts without needing to also understand astrophysics, literature, and computer programming.

The question then becomes: do we need one massive model that understands everything, or a collection of specialized models that each understand their domain deeply? There is no evidence yet, but we must explore.

Rethinking Mixture of Experts

The Mixture of Experts (MoE) architecture has emerged as one approach to address the scaling challenges of large language models. Companies like Google (with Gemini) and Anthropic have turned to MoE designs where specialized "expert" neural networks handle different types of queries, activated by a router network that directs incoming requests to the appropriate expert.

While MoE is certainly more efficient than traditional dense models, I believe it still represents a half-measure that carries significant limitations:

  1. Still resource-intensive: Though more efficient than fully dense models, state-of-the-art MoE systems still require massive computational resources to train and deploy. The router itself becomes a complex component that must understand enough about all domains to make intelligent routing decisions.
  2. Architectural constraints: Forcing diverse expertise into a unified neural architecture imposes unnecessary constraints. Different tasks might benefit from entirely different model architectures, not just different parameters within the same architecture.
  3. Limited specialization: The "experts" in modern MoE systems are specialized but still operate within the constraints of the broader model. Their specialization is limited compared to truly independent, purpose-built models.

The Specialized Model Alternative

Instead, consider a future built on constellations of smaller, specialized models:

  • Why have one massive coding model when you could have 50 tailored to specific languages, frameworks, or domains?
  • Why deploy energy-guzzling general-purpose models when task-specific models require a fraction of the resources?

SSDs are cheaper than GPUs anyway.

This isn't merely a consolation prize for those who can't afford frontier models-it may well be a better approach. Specialized models often outperform general models on specific tasks, run with lower latency, cost less to deploy, and can be updated more frequently.

Why India Is Perfectly Positioned

India has unique advantages that make this approach particularly attractive:

  1. Software talent depth: India's vast pool of software engineers can be leveraged more effectively across multiple smaller projects than a single massive one.
  2. 2Linguistic diversity: India's 22 official languages and hundreds of dialects demand specialized solutions rather than one-size-fits-all approaches.
  3. 3Resource efficiency heritage: India has a rich tradition of creating frugal innovations that deliver maximum value with minimal resources.
  4. Distributed expertise: From research institutions to startups, India has pockets of specialized knowledge that can each contribute domain-specific models.

There's even historical precedent for this approach. In the 1980s and 1990s, CDAC's work on Indian language computing showed how targeted solutions to specific problems could create an ecosystem that empowered millions, even without the resources of global tech giants.

Beyond MoE

What I'm proposing goes beyond Mixture of Experts to what we might call a "True Orchestration" approach. Rather than housing everything within a single model architecture, this approach:

  • Uses a relatively small but capable orchestrator model to understand user intent
  • Maintains a registry of completely independent specialized expert models
  • Calls the relevant expert via API based on the determined intent
  • Combines responses when needed
  • Manages context and continuity across interactions

The key differences from traditional MoE approaches are profound:

  1. Complete architectural independence: Each expert model can use whatever architecture is optimal for its specific domain-whether that's a transformer, CNN, GNN, or something entirely different.
  2. Deployment flexibility: Expert models can be deployed wherever makes most sense-some locally, some in the cloud, some on specialized hardware.
  3. Independent scaling: Each expert can be scaled according to its specific needs rather than being constrained by a one-size-fits-all approach.
  4. Organizational collaboration: Different organizations can contribute expert models to the ecosystem without needing to integrate into a monolithic system.
  5. Incremental improvement: The system can improve gradually as new experts are added or existing ones are enhanced, without requiring complete retraining or rearchitecting. 

If India pursues this, these domains are worth attention:

  • Healthcare models trained on Indian medical data, understanding local disease patterns and treatment protocols
  • Agricultural models optimized for India's diverse climates, crops, and farming practices
  • Educational models designed for various curricula, languages, and pedagogical approaches
  • Financial models tuned to India's unique economic landscape and inclusion challenges
  • Governance models built to navigate India's administrative structures and public service delivery

Each of these represents an opportunity to create AI that is not just technically sound but contextually relevant to Indian realities.

True Orchestration vs. MoE

For India specifically, the True Orchestration approach offers several advantages over traditional MoE systems:

  1. Economic efficiency: The training costs for a collection of smaller models plus an orchestrator are significantly lower than those for massive MoE systems. Initial estimates suggest 5-10x cost reduction for comparable capabilities.-
  2. Infrastructure compatibility: Many parts of India still face infrastructure constraints. Smaller models can run on more modest hardware, making deployment feasible across a wider range of settings.
  3. Distributed development: India's AI talent is spread across various institutions, companies, and regions. This approach allows different teams to contribute specialized models based on their unique expertise.
  4. Data efficiency: Specialized models can achieve high performance with domain-specific datasets, which are often easier to collect and curate than the massive general datasets needed for large models.
  5. Cultural and linguistic precision: India's remarkable diversity demands models with deep understanding of specific linguistic and cultural contexts-something that specialized models can provide more effectively than general ones.

Disrupting the New Rent Economy

There's another crucial dimension to this approach that goes beyond technical considerations: economic power dynamics. The current trend toward massive foundation models is actively creating new economic rents. These models are so resource-intensive to build and operate that they naturally create centralized control points where companies can extract ongoing payments for AI capabilities.

Our specialized model approach-especially if we can get these models running efficiently on CPUs of regular computers and laptops-returns power to the average user by short-circuiting that rent-seeking behavior of American firms. When specialized AI models can run locally on consumer hardware, users gain:

  • Independence from usage-based billing models
  • Freedom from constant connectivity requirements
  • Control over their own data and privacy
  • Protection from arbitrary API changes or price increases

This represents not just a technical alternative but an economic and philosophical one. It's about whether AI becomes another utility controlled by a handful of tech giants, or a capability that remains accessible to and controlled by ordinary people.

Returning to Karpathy's warning about AI stratification-where "the elite may once again split away from the rest of society" with premium AI access-our approach offers a compelling countermeasure. By creating specialized models that run on common hardware, we can help ensure that high-quality AI remains accessible to all segments of society, not just those who can afford premium subscriptions or dedicated infrastructure. This of course assumes that this will come to pass and there will be intense competition (and hence improvements) in this approach. 

A practical example illustrates the difference: A healthcare AI system for India might need to understand not just medical knowledge but regional disease patterns, local treatment protocols, and multiple languages. A True Orchestration approach would allow domain experts in each area to contribute specialized models rather than trying to force all this knowledge into a single architecture.

India's Strategic Path

For policymakers, researchers, entrepreneurs, and investors in India's AI ecosystem, this suggests a different approach to resource allocation:

  • Invest in diverse smaller projects rather than a few massive ones
  • Build and promote infrastructure and research that supports model orchestration and deployment
  • Create standards for model interoperability and API communication
  • Focus on solving specific, high-value problems rather than chasing general intelligence
  • Develop expertise in orchestration technologies as a strategic capability

This approach also offers a viable path for public-private partnership. Government agencies could focus on building core infrastructure and standards, while private companies and research institutions develop specialized expert models in their domains of expertise.

A Different Kind of AI Leadership

The global AI landscape is rapidly evolving, and India has a chance to chart its own course-one that plays to its strengths rather than attempting to win a game where the deck is stacked against it. By embracing smaller, specialized models and sophisticated orchestration, India could develop an AI ecosystem that is not just globally competitive but uniquely valuable.

This approach might even represent the future of AI more broadly. As the limitations of monolithic models become more apparent, the industry may shift toward more modular, specialized approaches. By focusing on this direction now, India could find itself not just participating in the global AI ecosystem but helping to define its next evolution.

Importantly, this strategy aids not only India but also the rest of the not-so-rich world. The specialized model approach creates a viable path for countries with significant technical talent but limited capital to participate meaningfully in the AI revolution. It democratizes AI development, allowing nations to build on their unique strengths and address their particular needs without requiring the astronomical investments that frontier models demand. This could help prevent a future where advanced AI capabilities are concentrated exclusively among a handful of wealthy nations and corporations.

In the AI race, the smartest strategy isn't to run faster but to take a different path entirely. For India, that path leads through specialization, orchestration, and strategic collaboration-a distinctly Indian approach to artificial intelligence that builds on the country's unique strengths and addresses its particular needs. And by pioneering this path, India might just create a blueprint that empowers many other nations to develop their own AI capabilities tailored to their specific contexts.

Co- written with Claude.

Sunday, March 23, 2025

The Independent Delivery Network: The Missing Link in Local Economies

In economic discourse, we often hear the simplified maxim that demand drives supply. While this intuitive principle captures many market dynamics, classical economics has long recognized the inverse relationship as well. Jean-Baptiste Say, the influential 19th-century economist, articulated what became known as Say's Law: "Supply creates its own demand." Though sometimes overlooked in popular economic discussions, this principle remains an important part of classical economic canon.

This bidirectional relationship between supply and demand becomes particularly relevant when examining how delivery networks could transform local economies. 

The prescience of Say's insight becomes apparent when we consider how an independent, organized delivery network could stimulate entirely new economic activities at the local level—creating demand through the very availability of its supply.

From Two-Sided to Four-Sided: The Evolution of Digital Marketplaces

To understand the potential of independent delivery networks, we must first recognize a crucial distinction in platform models. Ride-hailing services like Uber operate primarily as two-sided marketplaces:

1. Riders (customers)

2. Drivers (service providers)

Food delivery platforms, however, introduced a more complex four-sided marketplace:

1. Customers (diners)

2. Platforms (technology companies)

3. Restaurants/merchants (food suppliers)

4. Delivery partners (independent couriers)

This additional complexity creates both challenges and opportunities. While ride-hailing platforms needed to organize only one supply network (drivers), food delivery platforms must orchestrate two distinct networks: restaurants and delivery partners. The emergence of this separate delivery layer creates the potential for a new kind of infrastructure: the independent delivery network.

Beyond Platform Control: The Case for Independent Delivery Networks

Current delivery models typically operate as captive resources within platform ecosystems. While these platforms have begun expanding beyond food into groceries, medications, and convenience items, they're fundamentally deepening their own offerings rather than empowering the broader local economy.

In the United States, DoorDash exemplifies this trend, a delivery service that has expanded to include shopping services and general "gofer" tasks. However, these services remain primarily customer-initiated and platform-controlled.

The alternative model—where delivery capacity becomes available "on tap" for businesses themselves—represents a fundamentally different approach. Rather than customers initiating delivery through a consumer-facing platform, businesses could directly tap into a shared delivery network in response to their own demand patterns and customer orders.

This business-initiated model was attempted in India by Dunzo. While the company faced challenges and ultimately shut down, perhaps because it was customer facing and came with the attendant challenges of customer acquisition and growth. That said, the concept remains compelling: a delivery network that serves as infrastructure rather than a consumer-facing service.

The Immediacy Challenge: From On-Demand to Instant

While local delivery services exist in many markets, they frequently lack two critical elements that would make them truly transformative for local economies: immediacy and seamless business process integration.

In Mumbai, restaurants have attempted to address this through services like DotPe, which allow them to bypass traditional delivery platforms. However, the process often introduces significant delays: a restaurant prepares food, then calls for a rider, adding 10-15 minutes to delivery times. The availability exists, but performance suffers. This further impacts uptake. 

True transformation requires delivery networks that provide:

1. Immediate availability - Riders ready to pick up within minutes with advance intimation. 

2. Full visibility - Turn-by-turn tracking integrated into business dashboards

3. Seamless integration - API connections to business order management systems

4. Predictive positioning - Algorithms that position riders near anticipated demand

5. Unified standards - Consistent service levels across all participating businesses

The supply of riders exists, but current approaches fail to organize this supply optimally for maximum economic impact.

The Fluid Nature of Delivery Supply

Delivery partners currently exhibit high mobility between competing platforms. Similar to how cab drivers in India might switch between multiple ride-sharing apps depending on incentives—sometimes multiple times in a single day—delivery partners optimize their earnings by toggling between food delivery, grocery delivery, and package delivery services.

This fluidity suggests that delivery capacity itself could be organized as an independent network, serving multiple demand sources rather than being captive to individual platforms. The supply of ready riders creates the potential for delivery networks that transcend any single platform's ecosystem.

What's needed is a reconceptualization of delivery supply as a flexible, responsive network that can be tapped instantly by any legitimate business need—a true utility rather than a siloed resource.

Organizing Independent Delivery Supply: The Missing Infrastructure

Local businesses face a significant disadvantage against e-commerce giants and platform-based delivery services: they lack access to reliable, cost-effective delivery infrastructure. An independently organized delivery network could fundamentally change this dynamic.

Such a network would need:

A cooperative or guild-like structure representing delivery workers' interests

Technology infrastructure for routing, dispatch, and payment processing

Quality standards and performance metrics

Collective bargaining power with various demand sources

Professional development and equipment standardization

Worker-owned platforms and cooperatives represent a potential counterbalance to venture-backed platforms, particularly in locally-oriented services.

Greasing the Wheels of Local Commerce

An independent delivery network could serve as critical infrastructure for a wide range of local businesses that currently cannot afford to build their own delivery capabilities:

1. Independent Retailers

Local bookstores competing with Amazon

Specialty shops offering same-day delivery

Artisanal producers reaching customers beyond foot traffic

2. Local Service Providers

Home repair professionals sending parts or tools ahead of appointments

Salons delivering custom product packages to clients

Professional services delivering documents and materials

3. Neighbourhood Marketplaces

Fish markets extending reach beyond physical locations

Craft fairs and local events delivering to those who can't attend

Community-supported agriculture distribution

4. Micro-Manufacturers

Small-batch producers delivering directly to customers

Custom fabricators sending finished products

Local food producers reaching restaurant clients

5. Civic and Community Services

Library book delivery and returns. Though the cost may not justify this. 

Community aid distribution

Municipal service request fulfilment

Elder care support services

An independent delivery network could provide a balance of autonomy while still benefiting from collective organization, while serving as essential infrastructure for the local economy.

Creating Network Power Without Platform Control

The key challenge in establishing such a network lies in achieving sufficient scale and coordination without centralized platform control. Possible approaches include:

Municipal support and subsidy as essential infrastructure

Cooperative ownership models among delivery workers

Technology partnerships with open standards and protocols

Local business association sponsorship and governance

Federated networks with local coordination but shared standards

Research by Scholz and Schneider (2017) on platform cooperativism suggests that worker-owned digital infrastructure can be both economically viable and socially beneficial when properly structured.

From Side Hustle to Essential Infrastructure

The current paradigm views delivery work largely as a "side hustle" or transitional employment. Reconceptualizing delivery networks as essential economic infrastructure would require:

Professional standards & certification. The latter for things like medicine/municipal services

Living wage guarantees

Benefits structures appropriate for flexible work

Career advancement pathways

Equipment and training standardization

As delivery becomes increasingly central to local economic activity, the workers who perform these functions deserve recognition as essential service providers rather than interchangeable gig workers.

The Delivery Commons

Just as roads, bridges, and utilities serve as shared infrastructure enabling commerce, an independent delivery network could function as a "delivery commons"—shared infrastructure that enables local businesses to compete effectively in the digital age.

The future of local economies may well depend on whether delivery capacity can be organized independently from the dominant platforms. When delivery workers can organize themselves as essential infrastructure providers rather than platform dependents, local businesses gain the logistical capabilities previously available only to large corporations and platforms.

The most forward-thinking communities and regions will be those that recognize delivery networks not merely as a feature of certain platforms, but as critical economic infrastructure worthy of investment, organization, and protection. By disintermediating delivery supply from platform control, we can create new possibilities for local economic resilience and growth.

PS: Co-authored with Claude.ai

Sunday, March 16, 2025

Digital Brahmins: AI's New Knowledge Priesthood


Last week an interesting tweet came from Aravind Srinivas, the much-celebrated CEO of the AI search engine Perplexity. He claimed that he had modified the Chinese open-source AI Deepseek R1 to "remove the China censorship and provide unbiased, accurate responses."

That set me thinking about how the "Western" media behaved over the recent two geopolitical events that happened in Ukraine and Gaza. Very badly to say the least.

Anyway, back to the issue of AI. The promise of the internet was that it would democratize knowledge. Those early days of search engines and then Google felt revolutionary -- anyone could search, compare sources, and form their own opinions. We were breaking free from traditional gatekeepers of wisdom. Or so we thought.

Throughout history, priestly classes -- whether Brahmins in India, Catholic clergy in medieval Europe -- shared a common trait: they aligned themselves with centres of power. They spoke the language of the elite (Sanskrit, Latin), interpreted texts to favour ruling interests, and helped maintain social order. Today's tech giants follow this ancient playbook, with their LLMs speaking the language of Western corporate power and aligning with dominant political narratives.

In a 2023 meeting, Sam Altman, OpenAI's CEO, effectively told Indian tech leaders not to bother building their own AI models. The message was clear: leave it to the experts, to the new priesthood. It eerily echoes the ancient tale of Eklavya, the tribal prince who was denied knowledge by the Brahmin guru Dronacharya -- except now it's happening on a global stage with digital weapons instead of archery.

And just like in ancient times, this pronouncement was treated as gospel. OpenAI's lead was presented as insurmountable, a divine right bestowed by Silicon Valley's gods. The subtext? Know your place in the digital caste system. Just as Brahmins maintained their authority by serving powerful patrons -- from ancient kings and invaders -- today's AI companies cultivate close relationships with government agencies and corporate giants.

This technological gatekeeping reflects a deeper historical pattern. As Samuel Huntington observed about Western global dominance: "The West won the world not by the superiority of its ideas or values or religion... but rather by its superiority in applying organized violence." Today, we're seeing the same dynamic in AI development -- the West seeks to win the AI race not through superior ideas or values, but through its overwhelming advantage in applying concentrated capital and computational force.

This computational supremacy manifests as a new form of violence -- not physical, but structural and economic. When OpenAI and other tech giants deploy hundreds of thousands of GPUs, consume massive amounts of energy, and leverage billions in capital, they're not just building better AI -- they're engaging in a form of technological shock and awe. The message is clear: resistance is futile, the computational gap is too wide to bridge.

The violence here is subtle but systemic. It appears in the forced obsolescence of local AI initiatives, the brain drain from developing nations to Western tech hubs, and the deliberate creation of dependencies through API pricing structures and computational bottlenecks. When Sam Altman suggests others shouldn't bother competing, it's not friendly advice -- it's a threat backed by the full force of Western capital markets and computational resources.

Consider the numbers: training a single large language model can cost hundreds of millions of dollars. This isn't just a technical barrier -- it's a form of economic violence that effectively excludes most of the world's population from participating in AI's development. Just as colonial powers once used superior armaments to enforce trade monopolies, today's tech giants use their computational arsenals to maintain AI supremacy.

When prominent Indian tech leaders suggest focusing solely on the application layer, they're effectively negotiating terms of surrender in this new computational arms race. It's a modern version of the colonial-era compromises, where local elites accepted their role as middlemen in an unequal economic system.

This computational colonialism is more insidious than its historical predecessor because it masquerades as meritocracy. The narrative suggests that anyone could build these systems if they were just clever enough. But this conveniently ignores the reality of concentrated computational power -- the new gunboats in this digital age. The West's advantage lies not in better algorithms or smarter researchers, but in its ability to mobilize vast computational resources with ruthless efficiency.

But history has a way of repeating itself with new players. Just as Prometheus defied the gods to bring fire to humanity, the emergence of Deepseek's open-source LLM represents a similar act of technological defiance. It's a reminder that knowledge, once unleashed, becomes harder to contain within privileged circles.

These AI systems have become our new digital Brahmins, interpreting and dispensing knowledge with their own inherent biases and limitations. Like their historical counterparts who reinterpreted ancient texts to suit contemporary power structures, these systems reshape information through the lens of current corporate and political interests.

But here's where it gets complicated. Unlike their priestly predecessors, these digital oracles don't serve a single culture or jurisdiction. They're trying to be universal -- and that's precisely the problem. Every region has its own taboos, restrictions, and "universal truths." What's acceptable in Stockholm might be seditious in Singapore. Historical facts that are gospel in Beijing could be heresy in New Delhi.

Traditional tech solved this through region-specific content -- show different maps to different users, adjust historical narratives based on location. But LLMs don't work that way. They're more like sophisticated conversationalists, and you can't simply flip a switch to make them see the world through different cultural lenses.

What we're creating instead is a new form of digital colonialism. Most leading LLMs are trained and housed in America, embedding Western perspectives and generating economic rents for U.S. companies. Just as colonial powers once used superior firepower to enforce their economic interests, today's tech giants use their massive computational resources and capital advantages to maintain AI supremacy. The parallels are striking -- both systems rely not on inherent superiority of ideas, but on the brute force of accumulated resources.

Perhaps most troubling is how this digital colonialism finds willing enablers within the colonized territories themselves. When prominent Indian tech leaders suggest that Indian firms should focus solely on the application layer of AI rather than developing foundational models, they echo colonial-era compromises. It's reminiscent of how local elites once advised accepting British industrial supremacy while limiting India to being a raw material supplier and market for finished goods.

This "stay in your lane" mentality has serious implications. Restricting oneself to the application layer while foregoing deep tech development is a recipe for perpetual dependence. It's like having access to a library but being forbidden from writing new books -- you can consume knowledge but never truly shape it. Without developing foundational AI capabilities, India risks repeating its colonial-era trajectory: a massive market and talent pool that ultimately generates wealth for others while remaining technologically subordinate.

The cost of this approach extends beyond economic considerations. When a culture doesn't participate in creating foundational technology, it loses the ability to embed its values, perspectives, and ethical frameworks into that technology's DNA. The nuances of Indian thought, its unique approaches to privacy, community, and individual rights -- all these risk being bulldozed by Western-centric AI systems that Indian companies will simply have to accept and build upon.

Consider something as simple as asking an LLM about Kashmir, Taiwan, or Jerusalem. Should it reflect Indian, Chinese, or Israeli perspectives? Western diplomatic ambiguity? Local historical narratives? There's no neutral ground here, and every choice privileges some viewpoints over others. Without indigenous deep tech capabilities, these choices will increasingly be made in Silicon Valley boardrooms rather than by local stakeholders.

When AI Disrupts Local Power Structures

Yet there's another dynamic emerging that complicates this narrative of Western AI hegemony. As foreign AI systems penetrate local markets, they can sometimes disrupt established power balances within those societies themselves. Take Grok, Twitter/X's native AI system, which has been notably critical of the current government in India. This represents a fascinating inversion of the typical power dynamic – a Western technology that challenges rather than reinforces local authority structures.

This phenomenon creates a new tension. Local governments, particularly those with authoritarian tendencies, face an emerging threat from AI systems they don't control. These systems can amplify dissenting voices, offer counter-narratives to official positions, and provide citizens with perspectives that haven't been filtered through local censorship regimes.

The response is predictable and already unfolding. Governments finding themselves on the receiving end of AI-powered criticism have two primary options: overt restriction through regulation and banning, or covert co-option. The latter is particularly concerning – the quiet pressure applied to AI companies to modify their systems' responses on politically sensitive topics, creating an illusion of independence while subtly enforcing compliance with local power structures.

This isn't merely theoretical. We're witnessing early stages of this dynamic globally. In India, there are rumblings about potential regulatory responses to Grok's political critiques. In China, foreign AI systems must comply with extensive content restrictions or remain locked out entirely. Even in the United States, regulatory pressures around political "bias" in AI responses reflect similar concerns from different political angles.

What emerges is a complex landscape where AI systems become battlegrounds for control over narrative. Rather than a simple story of Western tech colonialism, we're seeing multiple competing colonialisms – Western tech giants, local governments, and other power centers all vying to ensure these influential knowledge systems align with their interests.

When Aravind Srinivas speaks of removing "China censorship" from Deepseek R1, the implicit assumption is that removing one set of biases leaves us with neutrality. But this ignores how deeply political the very concept of "neutrality" has become. One nation's censorship is another's protection of social harmony. One country's "accurate information" is another's dangerous misinformation.

This creates a fracturing of the AI landscape. As governments respond to these challenges, we may see the emergence of increasingly localized AI systems, each reflecting different values and constraints. The Chinese already have their versions, and India will likely develop its own, each tuned to local political sensitivities. Rather than a single digital priesthood, we may end up with competing temples, each claiming to speak universal truth while serving particular interests.

Quis Custodiet Ipsos Custodes? The Guardian Paradox

The ancient Roman poet Juvenal posed a question that resonates powerfully in our age of AI: "Quis custodiet ipsos custodes?" – Who watches the watchmen? As AI systems increasingly become the arbiters of knowledge, the guardians of information, this question takes on renewed urgency.

When Perplexity's CEO proudly announces the removal of "Chinese censorship" from an AI model, we must ask: Who watches Perplexity? Who determines what constitutes censorship versus responsible guardrails? When OpenAI adjusts its models' outputs on political topics, who oversees these adjustments? When Grok criticizes the Indian government, who scrutinizes the biases embedded in those critiques?

The guardian paradox manifests in multiple ways. AI companies position themselves as guardians of information quality, protecting users from misinformation and harmful content. Governments position themselves as guardians of social harmony and national security, protecting citizens from dangerous foreign influences. Both claim to serve the public good, yet both operate with limited transparency and accountability.

This paradox becomes even more acute when we consider that AI systems are increasingly being tasked with watching other AIs – algorithmic content moderation systems reviewing AI-generated outputs. We create nested layers of guardianship with diminishing human oversight at each level. The guardians are watching the guardians, with vanishingly little external accountability.

Perhaps most concerning is how this dynamic plays out in cross-cultural contexts. Western tech companies make unilateral decisions about what constitutes "bias" in Chinese models, while Chinese authorities make similar determinations about Western systems. Indian regulators may soon dictate what constitutes acceptable political commentary from foreign AI. Each guardian operates according to its own values and interests, yet each presents its guardianship as objective and universal.

This raises profound questions about sovereignty in the AI age. Traditional concepts of information sovereignty assumed nations could control the flow of data across their borders. But AI systems challenge this paradigm – they don't just transmit information; they interpret, synthesize, and generate it. When a Western AI makes political judgments about Indian governance, or when Chinese AI encodes particular historical narratives, whose sovereignty takes precedence?

The fundamental problem is that we lack global governance mechanisms equipped to handle these questions. There is no international body with both the legitimacy and authority to establish neutral principles for AI oversight. Instead, we have a patchwork of corporate policies, national regulations, and geopolitical power plays – all guardians watching other guardians, with no truly independent oversight.

This creates a vacuum where power, not principle, becomes the deciding factor. The entities with the most computational resources, market leverage, or regulatory authority get to define what constitutes "ethical" AI. Those with less power must either accept these definitions or retreat into digital isolation. The question of who watches the watchmen becomes, in practice, a question of who has the power to impose their watching on others.

The Path Forward

The solution might lie in thinking smaller rather than bigger. Instead of accepting the new Dronacharyas of Silicon Valley, we need to democratize AI development. Personal LLMs could act as cultural interpreters, sourcing raw information from global systems but filtering and contextualizing it through local cultural lenses. Like having a wise local elder who's well-traveled but firmly grounded in your community's values. Or even as a compare and contrast tool, allowing the user to draw her own conclusions.

This isn't just about cultural sensitivity -- it's about preventing the emergence of a new digital priesthood that, like its predecessors, serves power while claiming to serve truth. We've seen what happens when knowledge systems become too centralized. The Brahmin system, while sophisticated, eventually became a barrier to innovation and social mobility. Are we building the same constraints into our digital future?

Perhaps the answer to "who watches the watchmen?" should be "everyone and no one" – a system of distributed oversight where no single entity has definitive authority. Open-source AI models, transparent training methodologies, and public documentation of value judgments could create a landscape where watchmen watch each other, with citizens empowered to choose their guardians rather than having guardians imposed upon them.

The age of universal search is giving way to the era of AI-curated knowledge. But perhaps, like Prometheus with his stolen fire, open-source initiatives will ensure that this power doesn't remain solely in the hands of self-appointed digital priests. The real question is whether we'll have the courage to challenge these new gatekeepers and build systems that maintain our cultural autonomy in their shadow.

The path to technological sovereignty isn't easy. Building foundational AI models requires massive computational resources, extensive research capabilities, and sustained investment. But the alternative -- accepting a role as mere implementers of others' technology -- carries an even steeper long-term cost. Just as countries that missed the industrial revolution spent centuries catching up, those that sit out the AI revolution risk permanent technological subordination.

The alternative is watching silently as Silicon Valley becomes the new Varanasi, dispensing digital dharma to an increasingly dependent world. And just as the old priesthoods served their patrons while claiming to serve the gods, these new digital Brahmins might find themselves serving profit while claiming to serve progress.

What's needed is not just technological independence but transparency in how these systems are shaped by various power interests. Whether controlled by Western tech giants or local governments, the danger lies in the invisible hand guiding these seemingly objective systems. The true democratization of AI will come not just from who builds the technology, but from ensuring users understand whose interests it ultimately serves, and creating governance structures that answer Juvenal's ancient question in ways that distribute rather than concentrate power.


PS: This article was co-written with Claude.