๐ดWhat is red-teaming?
Red Teaming is roleplaying as an attacker. A practice dopted from the military into infosec and then info machine learning eval, in red teaming, humans try to get a system to fail. Humans are pretty creative, and usually up-to-date, and this works pretty fine.
Resources about red teaming:
Red Team Guide, https://redteam.guide/docs/guides
NVIDIA AI Red Team: An Introduction, https://developer.nvidia.com/blog/nvidia-ai-red-team-an-introduction/
One thing the human activity of red teaming doesnโt do is to scale. Itโs great for intelligence gathering, and as a source of generative material for creativity, but it doesnโt scale great. Human expertise is expensive, and good red-teamers are few and far between. Iโm not saying that many red teamers are bad โ simply that there arenโt many people who can do this well in the first place.
What if we could automate some of the basics?
Last updated