Security Incidents Will Drive Demand
Security incidents in AI agent systems will occur and create market pressure for isolation, similar to how browser security incidents drove sandboxing adoption.
The Assumption
This is the hidden foundation beneath “Agents Need Sandboxes”. We’re not betting on current pain—we’re betting on future pain. The browser sandboxing analogy only works if security incidents actually occur and drive demand.
Today, many developers run Claude Code, Cursor, and other AI agents directly on their machines without sandboxing. No high-profile security incidents have occurred yet. Our thesis requires this to change.
Evidence
Supporting signals:
- Security researchers actively exploring prompt injection → code execution chains
- Academic papers documenting theoretical attack vectors
- Historical pattern: browsers, Flash, Java all required incidents before sandboxing
- Attack surface expanding as agents gain more capabilities
Counter-signals:
- No major incidents despite widespread agent adoption
- Agent code is typically user-approved, unlike browser code from untrusted sites
- Attack surface may remain theoretical
- Users may accept risk in exchange for convenience
What Would Prove This Wrong
- 24 months pass with no significant agent security incidents
- Agents deployed widely without isolation and nothing bad happens
- Security community consensus emerges that agent risks are overstated
- Users consistently choose convenience over security
Impact If Wrong
If security incidents don’t occur, the urgency for sandboxing evaporates. SmartBoxes becomes a “nice to have” rather than a “must have”. We’d need to pivot the value proposition from security to convenience, developer experience, or operational simplicity.
Testing Plan
Monitoring:
- CVE databases for agent-related vulnerabilities
- Security research publications and conference talks
- Hacker News, Reddit discussions of agent security
- Enterprise security team concerns in customer discovery
Timeline: 12 months to initial signal
Kill criteria: If 24 months pass with no incidents and no enterprise security concerns, the foundation of our thesis is shaky.
Related
This is a bedrock assumption that “Agents Need Sandboxes” implicitly depends on.
Enables:
- Agents Need Sandboxes — the parent assumption
Assumption
Security incidents in AI agent systems will occur and create market pressure for isolation, similar to how browser security incidents drove sandboxing adoption.
Enables
If this assumption is true, these become relevant:
- Agents Need Sandboxes — 🏛️ ⚪ 70%
How To Test
Monitor security research, CVE databases, and news for agent-related incidents. Track enterprise security team concerns.
Validation Criteria
This assumption is validated if:
- Documented prompt injection to code execution chains in the wild
- Enterprise security teams blocking agent deployment citing isolation
- Insurance or compliance requirements emerge for agent isolation
Invalidation Criteria
This assumption is invalidated if:
- 24 months pass with no significant agent security incidents
- Agents deployed widely without isolation and no problems emerge
- Security community consensus that agent risks are overstated
Dependent Products
If this assumption is wrong, these products are affected: