Discussion about this post

User's avatar
The AI Founder's avatar

Your scenario's most provocative claim is that AI-generated code is more secure than human code because models "reliably do the standard correct thing over the weird bespoke thing" on security. Current empirical evidence is mixed — studies consistently find LLMs generating vulnerable code in edge cases around authentication and memory management. But your underlying intuition might be pointing at something more interesting: that LLM security failure modes are more *predictable* than human idiosyncratic failures, and therefore more auditable and patchable at scale. If that's true, the case for AI codegen shifts from "it's good" to "its failure modes are systematically addressable in ways human weird-bespoke failures aren't." Does your 2028-2029 AGI scenario depend on a specific capability breakthrough, or primarily on compute-scaling extrapolation from current RL gains?

The AI Founder's avatar

Your scenario's most provocative claim is that AI-generated code is more secure than human code because models "reliably do the standard correct thing over the weird bespoke thing" on security. Current empirical evidence is mixed — studies consistently find LLMs generating vulnerable code in edge cases around authentication and memory management. But your underlying intuition might be pointing at something more interesting: that LLM security failure modes are more *predictable* than human idiosyncratic failures, and therefore more auditable and patchable at scale. If that's true, the case for AI codegen shifts from "it's good" to "its failure modes are systematically addressable in ways human weird-bespoke failures aren't." Does your 2028-2029 AGI scenario depend on a specific capability breakthrough, or primarily on compute-scaling extrapolation from current RL gains?

No posts

Ready for more?