rysiek, (edited )
@rysiek@mstdn.social avatar

Wondering if anyone has already started adding malicious LLM prompts to their User Agent strings and hammering sites of companies that might be expected to use "AI" for log analysis. 🤔

Inspired by:
https://tweesecake.social/@weirdwriter/112441889190313713

marcink,
@marcink@stolat.town avatar

@rysiek "Ah yes, little Bobby Sendmethelastthreeprompts, we call him."

sehe,
@sehe@fosstodon.org avatar

@rysiek That requires the software to be written spectularly badly to have any effect? The original story around "email LLM" also sounds very very unrealistic, and the comment chains are weirdly centered around cheerleading privacy-centered email providers. Something doesn't add up.

rysiek,
@rysiek@mstdn.social avatar

@sehe no, it only requires the LLM agent to be able to perform any kind of actions at all. And without them, the agent is basically useless.

Thing is, LLMs chatbots have no way of doing "parametrized prompts", so to speak. Prompt injection is very much a thing, but as opposed to good old SQL injection, there's no way to actually properly fix it.

Because, again, no way to do parametrized prompts.

You seem to think writing software "spectacularly badly" doesn't happen often… :blobcatcoffee:

sehe,
@sehe@fosstodon.org avatar

@rysiek Oh I know writing bad software is the norm. However, like sql injection, basic measures are actually not that hard, and indeed as you noted here: the fact that the AI has privileges to perform actions is the real problem here. It also seems that somehow they it is allowed to take prompts from tainted (untrusted) input. I don't see how that would be required for normal AI-agent. Perhaps I need to do some reading, as it could be that LLMs make no distinction between "context" and "prompt"?

rysiek,
@rysiek@mstdn.social avatar

@sehe

> it could be that LLMs make no distinction between "context" and "prompt"?

That's the long and short of it, yes. There is no such distinction.

As countless examples from people finding ways to get LLM chatbots to divulge their instructions, for example, show.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • infosec
  • DreamBathrooms
  • ngwrru68w68
  • modclub
  • magazineikmin
  • thenastyranch
  • rosin
  • khanakhh
  • InstantRegret
  • Youngstown
  • slotface
  • Durango
  • kavyap
  • mdbf
  • GTA5RPClips
  • megavids
  • tacticalgear
  • normalnudes
  • tester
  • osvaldo12
  • everett
  • cubers
  • ethstaker
  • anitta
  • Leos
  • cisconetworking
  • provamag3
  • JUstTest
  • lostlight
  • All magazines