Malicious prompt injections to manipulate generative artificial intelligence (GenAI) large language models (LLMs) are being wrongly compared to classical SQL injection attacks. In reality, prompt ...