Welcome to DU! The truly grassroots left-of-center political community where regular people, not algorithms, drive the discussions and set the standards. Join the community: Create a free account Support DU (and get rid of ads!): Become a Star Member Latest Breaking News Editorials & Other Articles General Discussion The DU Lounge All Forums Issue Forums Culture Forums Alliance Forums Region Forums Support Forums Help & Search

highplainsdem

(60,076 posts)
Fri Jan 9, 2026, 07:05 PM Friday

ChatGPT falls to new data-pilfering attack as a vicious cycle in AI continues

https://arstechnica.com/security/2026/01/chatgpt-falls-to-new-data-pilfering-attack-as-a-vicious-cycle-in-ai-continues/

STILL LEAKY AFTER ALL THESE YEARS
ChatGPT falls to new data-pilfering attack as a vicious cycle in AI continues
Will LLMs ever be able to stamp out the root cause of these attacks? Possibly not.

Dan Goodin – Jan 8, 2026 8:00 AM


There’s a well-worn pattern in the development of AI chatbots. Researchers discover a vulnerability and exploit it to do something bad. The platform introduces a guardrail that stops the attack from working. Then, researchers devise a simple tweak that once again imperils chatbot users.

The reason more often than not is that AI is so inherently designed to comply with user requests that the guardrails are reactive and ad hoc, meaning they are built to foreclose a specific attack technique rather than the broader class of vulnerabilities that make it possible. It’s tantamount to putting a new highway guardrail in place in response to a recent crash of a compact car but failing to safeguard larger types of vehicles.

-snip-

As is the case with a vast number of other LLM vulnerabilities, the root cause is the inability to distinguish valid instructions in prompts from users and those embedded into emails or other documents that anyone—including attackers—can send to the target. When the user configures the AI agent to summarize an email, the LLM interprets instructions incorporated into a message as a valid prompt.

AI developers have so far been unable to devise a means for LLMs to distinguish between the sources of the directives. As a result, platforms must resort to blocking specific attacks. Developers remain unable to reliably close this class of vulnerability, known as indirect prompt injection, or simply prompt injection.

-snip-



Much more at the link. No paywall.

One of many reminders over the years that LLMs aren't very secure.

And this latest news about that is especially timely, given that OpenAI is now urging people to trust them with their confidential medical data: https://www.democraticunderground.com/100220920563
3 replies = new reply since forum marked as read
Highlight: NoneDon't highlight anything 5 newestHighlight 5 most recent replies
ChatGPT falls to new data-pilfering attack as a vicious cycle in AI continues (Original Post) highplainsdem Friday OP
AI is "intelligence" without morality... Wounded Bear Friday #1
Remember years back when some tv advertisement used "Alexa" or some vocal sounds close enough to the word... Hugin Friday #2
So much effort for humans to do less work OC375 Friday #3

Wounded Bear

(63,857 posts)
1. AI is "intelligence" without morality...
Fri Jan 9, 2026, 07:12 PM
Friday

Intelligence without the capacity to distinguish right from wrong is pure evil. I'm not a religious person. I believe that morality is based not on any religion but is based solely on a recognition that actions and words can build a better world or can destroy civilization.

Hugin

(37,448 posts)
2. Remember years back when some tv advertisement used "Alexa" or some vocal sounds close enough to the word...
Fri Jan 9, 2026, 07:21 PM
Friday

And it activated user’s Amazon Home systems to do things?

Yeah, this is the same thing.

The chatbots are always parsing and interpreting. Because that’s how the probability engines in them work.

Latest Discussions»General Discussion»ChatGPT falls to new data...