prompt injection

A malicious method that crafts model inputs to override instructions, bypass safeguards, or coax sensitive outputs from a language model.