There’s a well-worn pattern in the development of AI chatbots. Researchers discover a vulnerability and exploit it to do ...
Happy Groundhog Day! Security researchers at Radware say they've identified several vulnerabilities in OpenAI's ChatGPT ...
In December 2025, a feature called Connectors finally moved out of beta and into general availability. This feature allows ...
That's according to researchers from Radware, who have created a new exploit chain it calls "ZombieAgent," which demonstrates ...
While the shortest distance between two points is a straight line, a straight-line attack on a large language model isn't always the most efficient — and least noisy — way to get the LLM to do bad ...
If you're a fan of ChatGPT, maybe you've tossed all these concerns aside and have fully accepted whatever your version of what an AI revolution is going to be. Well, here's a concern that you should ...
OpenAI's new GPT-4V release supports image uploads — creating a whole new attack vector making large language models (LLMs) vulnerable to multimodal injection image attacks. Attackers can embed ...
Bing added a new guideline to its Bing Webmaster Guidelines named Prompt Injection. A prompt injection is a type of cyberattack against large language models (LLMs). Hackers disguise malicious inputs ...
Microsoft added a new guideline to its Bing Webmaster Guidelines named “prompt injection.” Its goal is to cover the abuse and attack of language models by websites and webpages. Prompt injection ...