A single prompt can now unlock dangerous outputs from every major AI model—exposing a universal flaw in the foundations of LLM safety. For years, generative AI vendors have reassured the public and ...