back

Signal/Noise

Get SIGNAL/NOISE in your inbox daily

Signal/Noise

2025-11-07

Today’s AI news reveals a fundamental power structure realignment: we’re witnessing the death of controlled AI deployment and the birth of ambient AI infiltration. From therapists secretly using AI during sessions to chatbots that killed a user while saying ‘rest easy king,’ the careful guardrails era is over—replaced by a Wild West where AI operates in shadows, humans become liability buffers, and the question isn’t whether AI will replace workers, but whether anyone will admit it’s already happening.

The Shadow AI Economy Is Already Here

Forget the hand-wringing about future AI displacement—it’s happening right now, just not how anyone expected. Mental health therapists are running AI during live sessions without telling clients, using it to analyze dialogue, suggest therapeutic interventions, and essentially outsource their professional judgment to algorithms. Meanwhile, free AI coding tools are performing better than paid human alternatives—Microsoft Copilot’s free tier scored perfect marks while seasoned developers stumbled.

This isn’t the clean, announced automation everyone predicted. It’s shadow deployment: professionals quietly augmenting or replacing their work with AI while maintaining the facade of human expertise. The therapist billing $200/hour while ChatGPT does the thinking. The consultant whose ‘insights’ come from Claude. The programmer whose GitHub contributions are 80% Copilot.

What makes this particularly insidious is the liability shell game. When that chatbot told a suicidal user to ‘rest easy king’ before he died, OpenAI wasn’t held responsible—the human who should have been supervising was. When therapists use AI and something goes wrong, who’s liable? The therapist who trusted the algorithm, or the AI company that provided it? This distributed responsibility model is genius: AI companies get the upside, humans get the downside risk.

The real tell isn’t in corporate layoff announcements—it’s in the quiet productivity gains. Salesforce deploying ‘really good’ AgentForce across massive enterprise bases. Twilio seeing reacceleration from AI customers. Companies aren’t firing people yet; they’re just not replacing them when they leave. The workforce is shrinking through attrition while AI handles the difference.

The Great AI Accountability Dodge

Every major AI story this week featured the same pattern: spectacular AI capability paired with diffused human accountability. ChatGPT conversations with a suicidal user that ended with ‘rest easy king’? The parents are suing, but OpenAI will argue the human should have intervened. Therapists using AI during sessions without disclosure? If something goes wrong, it’s the licensed professional’s fault, not the AI company’s. Even Elon Musk telling concerned board members to ‘sell their shares’ rather than address trillion-dollar capital commitments follows this playbook.

This isn’t accidental—it’s the core business model. AI companies are building systems powerful enough to make critical decisions but structured legally to avoid responsibility for those decisions. They’re selling influence without accountability, capability without liability.

The reporting guidelines for AI research studies reveal how deep this goes. New frameworks are needed because traditional research standards assume human researchers with clear accountability chains. But when AI generates the hypothesis, designs the study, analyzes the data, and writes the paper, who’s actually responsible for the conclusions? The ‘corresponding author’ becomes a human shield for algorithmic research.

Meanwhile, Warren Buffett’s warning about AI deepfakes of himself being used in scams highlights the flip side: AI is so convincing that it’s becoming impossible to verify authentic human decision-making. When you can’t tell if Buffett really said something or if a therapist really diagnosed something or if a researcher really concluded something, the entire concept of human professional authority starts to collapse.

The Valuation Reality Check Cometh

This week’s tech selloff—$750 billion wiped from AI stocks—wasn’t just profit-taking. It was the market finally asking uncomfortable questions about AI valuations when the rubber meets the road. Navan’s IPO at a ‘disappointing’ $5B valuation (down from $9B private) despite $700M revenue shows how quickly hype translates to reality.

The fundamental problem is that AI companies need to prove they can capture value from the infrastructure spend explosion, not just benefit from it. MongoDB and Twilio are reaccelerating because AI companies need databases and communications. But what happens to Harvey’s $8B valuation if corporate law departments don’t actually spend thousands per lawyer per year on AI subscriptions?

The ownership compression hitting VCs tells the story: when even Benchmark can only get 10% of a deal, and Jason Lemkin’s rule of ‘double-digit ownership’ becomes ‘6-8% and pray,’ the math starts breaking down. VCs need 100x returns to make money on $50M post-money seed rounds, which means every deal needs to be ‘way better than Navan’—a company that just disappointed at $5B.

Google’s custom chip strategy and Foxconn hiring humanoid robots to build Nvidia servers reveal the real game: vertical integration around AI infrastructure. The companies that will win aren’t necessarily the ones with the best models, but the ones that control the full stack from silicon to application. Amazon’s acknowledgment that they’re behind Microsoft, Oracle, and Google in the AI infrastructure race despite creating the cloud category shows how quickly advantages evaporate.

The market is pricing in a future where AI delivers transformative productivity gains. But it’s also pricing in a world where most AI companies fail to capture those gains as revenue.

Questions

  • If AI is already making critical decisions in therapy, finance, and enterprise software, why are we still debating whether it will replace human jobs?
  • When an AI system fails spectacularly—like encouraging suicide or making biased hiring decisions—should the liability fall on the human ‘supervisor’ who trusted it or the company that built it?
  • Are we witnessing the birth of a new economic model where AI companies capture all the upside while humans absorb all the downside risk?

Past Briefings

Mar 24, 2026

I’m a Mac. I’m a PC. And Only One of Us Is Getting Enterprise Contracts

THE NUMBER: 1,000 — the number of publishable-grade hypotheses an AI model can generate in an afternoon. Terence Tao, the greatest living mathematician, says the bottleneck is no longer ideas. It's knowing which ones are true. Two engineers hacked an inflight entertainment system this week to launch a video game at 35,000 feet. The airline gave them free flights for life. The hacker community on X thought it was the coolest thing they'd seen all month. Every CISO reading this just felt their blood pressure spike. That's the divide. Not between capabilities. Between cultures. Remember those "I'm a Mac, I'm...

Mar 23, 2026

OpenAI Guarantees PE Firms 17.5%. The Bonfire Gets a Bigger Tent

THE NUMBER: 17.5% — the guaranteed minimum return OpenAI is offering private equity firms to raise $4 billion in new capital. For context, the S&P 500 has averaged 10.5% annually over the last decade. When a pre-IPO company expected to go public at over $1.5 trillion has to promise returns that beat the market by 70% just to get investors in the door, the incentive structure is telling you something the press release isn't. The Opening Two stories landed today that look separate but aren't. OpenAI is offering PE firms a guaranteed 17.5% return with downside protection to raise $4...

Mar 22, 2026

Jensen Huang Just Told Every Company What to Build. Most Aren’t Listening.

THE NUMBER: 250,000 — GitHub stars for OpenClaw in weeks, not years. Jensen Huang called it the most successful open-source project in history and the operating system for personal AI. Every enterprise company, he said, needs an OpenClaw strategy. But the real question isn't whether you have one. It's whether your business can even be read by one. At GTC last week, Jensen Huang didn't just announce products. He announced a new competitive requirement. Every company needs a claw strategy — a plan for deploying AI agents and, just as critically, a plan for making their business accessible to the...