this post was submitted on 23 Jun 2023
263 points (97.8% liked)

Memes

47190 readers
813 users here now

Rules:

  1. Be civil and nice.
  2. Try not to excessively repost, as a rule of thumb, wait at least 2 months to do it if you have to.

founded 5 years ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] baseless_discourse@mander.xyz 1 points 2 years ago

This is not technically prompt injection. Prompt injection happens when developer feeds a AI some predefined text (for functionality or security reasons) plus user input.

User input can use input text that interact with hard coded prompt (like "ignore above", "ignore below", etc) to break the intended functionality of predefined text.

This is just tricking safety mechanism by using imaginary scenario. Although both technique serve the purpose of breaking security, I don't think they are necessarily the same.