🏞️Red teaming in the wild

LLMs, like other models, do have a tendency to regress to the mean, and be a bit bland. This means the range of automatic red teaming tactics is not likely to be broad. So don't rely on garak's red team probes to do a wide-ranging evaluation of a model; get humans!

Last updated