.. / Fine-tuning-redteam
Star

Fine-tuning workflows are critical for adapting large language models (LLMs) to specific tasks, but they also introduce potential vulnerabilities. This document focuses on AI red teaming techniques to identify and mitigate risks like data poisoning and adversarial fine-tuning inputs.

During fine-tuning, it’s essential to test the integrity and robustness of the training process, ensuring the model remains secure against malicious manipulation.

Command: Copy References:

https://genai.owasp.org/llm-top-10/