The analogy is that AI is suppose to be able to do _What humans do_ but better.
But you also want AI to be more secure. To make it more secure, you'll have to prevent the user from doing things _they already do_.
Which is impossible. The current LLM AI/Agent race is a non-deterministic GIGO and will never be secure because it's fundamentally about mimicing humans who are absolutely not secure.
But you also want AI to be more secure. To make it more secure, you'll have to prevent the user from doing things _they already do_.
Which is impossible. The current LLM AI/Agent race is a non-deterministic GIGO and will never be secure because it's fundamentally about mimicing humans who are absolutely not secure.