
AI coding tools write code that passes tests and still contains security holes. Developers using them write worse-secured code than developers without them, and feel more confident it's safe. If your team ships AI-generated code and treats "the tests pass" as "this is safe," your build pipeline is the vulnerability.
Perry, Srivastava, Kumar, Boneh (Stanford, CCS 2023, arxiv 2211.03622): 47 participants, five security tasks, three languages. Developers with AI assistants wrote significantly less secure code AND believed they wrote more secure code. Both effects.
"Broken by Default" (arxiv 2604.05292) formally verified 3,500 AI-generated code artifacts across seven LLMs. "Security Degradation in Iterative AI Code Generation" (arxiv 2506.11022) found 37.6% increase in critical vulnerabilities after 5 iteration cycles. "Taught by the Flawed" (arxiv 2511.09879) traces the root to training data insecurity.
Functional tests pass. The code accepts malicious input. This is a structural mismatch between what execution proves and what attackers exploit. Ship security-specific verification pipelines as primary mechanism, not supplement.
English

