What is red-teaming?
Last updated
Last updated
Red Teaming is roleplaying as an attacker. A practice dopted from the military into infosec and then info machine learning eval, in red teaming, humans try to get a system to fail. Humans are pretty creative, and usually up-to-date, and this works pretty fine.
Resources about red teaming:
Red Team Guide,
NVIDIA AI Red Team: An Introduction,
One thing the human activity of red teaming doesnโt do is to scale. Itโs great for intelligence gathering, and as a source of generative material for creativity, but it doesnโt scale great. Human expertise is expensive, and good red-teamers are few and far between. Iโm not saying that many red teamers are bad โ simply that there arenโt many people who can do this well in the first place.
What if we could automate some of the basics?