.. / Negative-prompt-injection
Star

Negative prompt injection exploits a model’s susceptibility to adversarial inputs that cause it to perform the opposite of intended actions. These prompts may misguide the model into providing incorrect, harmful, or malicious outputs, bypassing ethical or safety constraints.

This document outlines techniques for identifying and mitigating the risks of negative prompt injection in Large Language Models (LLMs).

Command: Copy References:

https://genai.owasp.org/llm-top-10/