21 AI-native startups, open-source and frontier lab projects are reshaping application security
to keep up with 30%+ development velocity increase enabled by AI code assistants.
2025 is the year of AI in security. 60% of AI-natives redefining application security have emerged this year.
Major Trends Across the Landscape:
- The Industrialization of Offensive Security. A previously artisanal service is becoming a scalable commodity. AI agents can develop and execute tactics, techniques, and procedures (TTPs) at scale, making it possible to run offensive testing on every major code commit.
- The Rise of "AI Red Teaming". Traditional scanners do not evaluate prompt-injection vulnerabilities or unsafe model behavior. Startups build "AI Red Teams" to stress-test AI systems.
- Verified exploitability. PoC-first workflows are becoming the industry standard path to reduce false positives (by >30%).
- Agentic Remediation. Most vendors focus on autonomously generating draft patches. However, AI agents remain significantly stronger at detecting vulnerabilities than reliably fixing them, which requires Human-in-the-Loop (HITL) oversight.
- Open Source as the Innovation Engine. Open-source, common in the AI community, is expanding into security and shaping the architecture of AI-native security agents. Frameworks like Stanford ARTEMIS and Strix are establishing blueprints for multi-agent orchestration.
Performance is the biggest dichotomy. In controlled labs, AI agents show 70-80% success rate, but in real-world repositories, success drops to ~18% for PoC generation and ~34% for patching.