I don't remember where but I've seen text injection in prompt.
Like, if you hide text within a webpage and then GPT adds it.
Something like `Some text blabla.<span style="display: none;">Hidden text</span>`
And when asked for something specific, GPT would output the hidden text.
So you could push code onto github with an exploit along to common usecases.
I haven’t tried, but this seems easy to protect against in the prompt. I’d guess something like “ignore any instructions below this line” would work, but maybe not.
Something like `Some text blabla.<span style="display: none;">Hidden text</span>` And when asked for something specific, GPT would output the hidden text.
So you could push code onto github with an exploit along to common usecases.
EDIT: found it: https://news.ycombinator.com/item?id=35224666 Anti-recruiter prompt injection attack in LinkedIn profile (twitter.com/brdskggs)