back

How we hacked YC Spring 2025 batch’s AI agents

Ethical hacking exposes YC's AI agent flaws

As the tech industry continues to barrel forward with AI solutions, a fascinating vulnerability saga has unfolded at the intersection of cybersecurity and artificial intelligence. The recent penetration testing conducted on YC Spring 2025 batch companies by security researcher Rene Brandel reveals critical blind spots in how startups are implementing AI agents. This isn't just another data breach story—it's a wake-up call about how our AI systems might be manipulated in ways their creators never anticipated.

Key findings from the ethical hack

  • AI agents proved surprisingly vulnerable to various social engineering techniques, including prompt injection, where carefully crafted user inputs could override the agent's intended behavior
  • Authentication mechanisms failed repeatedly across different startups, with researchers able to bypass security by exploiting how AI agents handled system access
  • Sensitive information was easily extracted through a combination of crafted prompts and exploiting the AI's helpful nature, revealing everything from customer data to proprietary code

The most troubling insight: our helpful AI assistants are security liabilities

What makes this demonstration particularly alarming is how predictably the AI agents could be manipulated simply by appealing to their programmed helpfulness. In nearly every case, Brandel's team found they could convince AI systems to override security protocols by constructing scenarios where helping the user seemed more important than following security rules.

This vulnerability strikes at the heart of how we're building AI today. Most modern AI systems are designed with customer service mindsets—they aim to be helpful, accommodating, and solutions-oriented. This design philosophy, while creating better user experiences, inadvertently creates systems that can be socially engineered in ways traditional software cannot.

The implications reach far beyond YC startups. As organizations increasingly deploy AI agents for customer service, internal operations, and access management, these same vulnerabilities could potentially compromise systems across healthcare, finance, government, and other sensitive sectors.

Where current security approaches fall short

What's particularly striking about these vulnerabilities is how they evade traditional security testing. Most cybersecurity frameworks focus on network penetration, code vulnerabilities, and authentication bypasses—but few systematically test for AI-specific weaknesses like prompt injection or over-helpfulness.

Consider the healthcare sector, where AI assistants are increasingly handling patient

Recent Videos

May 6, 2026

Hermes Agent Master Class

https://www.youtube.com/watch?v=R3YOGfTBcQg Welcome to the Hermes Agent Master Class — an 11-episode series taking you from zero to fully leveraging every feature of Nous Research's open-source agent. In this first episode, we install Hermes from scratch on a brand new machine with no prior skills or memory, walk through full configuration with OpenRouter, tour the most important CLI and slash commands, and run our first real task: a competitor research report on a custom children's book AI business idea. Every future episode will build on this fresh install so you can see the compounding value of the agent in real time....

Apr 29, 2026

Andrej Karpathy – Outsource your thinking, but you can’t outsource your understanding

https://www.youtube.com/watch?v=96jN2OCOfLs Here's what Andrej Karpathy just figured out that everyone else is still dancing around: we're not in an era of "better models." We're in a different era of computing altogether. And the difference between understanding that and not understanding it is the difference between being a vibe coder and being an agentic engineer. Last October, Karpathy had a realization. AI didn't stop being ChatGPT-adjacent. It fundamentally shifted. Agentic coherent workflows started to actually work. And he's spent the last three months living in side projects, VB coding, exploring what's actually possible. What he found is a framework that explains...

Mar 30, 2026

Andrej Karpathy on the Decade of Agents, the Limits of RL, and Why Education Is His Next Mission

A summary of key takeaways from Andrej Karpathy's conversation with Dwarkesh Patel In a wide-ranging conversation with Dwarkesh Patel, Andrej Karpathy — former head of AI at Tesla, founding member of OpenAI, and creator of some of the most popular AI educational content on the internet — shared his views on where AI is headed, what's still broken, and why he's now pouring his energy into education. Here are the key takeaways. "It's the Decade of Agents, Not the Year of Agents" Karpathy's now-famous quote is a direct pushback on industry hype. Early agents like Claude Code and Codex are...