Prompt Injection
Prompt Injection is a type of security attack where an adversary manipulates the input to a text generation system, like a language model, to produce unintended or harmful outputs. This occurs when model inputs are altered with external, sometimes malicious, data that leads the system to act against its original instructions. For instance, altering chatbot instructions to produce inappropriate responses is a typical example of prompt injection.