Red Teaming
& Evaluation

Testing GenAI systems through adversarial red teaming methods.

This project establishes comprehensive AI Red Teaming and evaluation guidelines for Large Language Models (LLMs), addressing security vulnerabilities, bias, and user trust. By collaborating with partners and leveraging real-world testing, the initiative will provide a standardized methodology for AI Red Teaming, including benchmarks, tools, and frameworks to boost cybersecurity defenses.

Resource Links:

What’s New

Get Started

Quick access to meetings and collaboration groups
Weekly

Monday

9:30 AM PDT
10:30 AM PDT
Open Meeting – AI Threat Intelligence

Weekly initiative meeting.

Add to Calendar

Initiative Leads

Sonu Kumar

Initiative Leaders

Jason Ross

Core Team MemberInitiative Leaders

Community of Contributors

Explore a global network of volunteers improving evaluations, patterns, and defenses for autonomous systems.
Scroll to Top