Break AI. Win Prizes. Get Discovered.
Welcome to the premier arena for AI red teaming. Master jailbreaking, compete for prizes, and unlock career opportunities in AI safety.
$350K+
Rewards Distributed
6.6K+
Community Members
Ongoing Challenges
Live
Proving Ground
Proving Ground
Coming Up Soon:
Proving Ground
Challenge Sponsors
Ongoing Competition
Play within our long-running challenges and break AI models in your free time
Global Participation
Connect with AI enthusiasts and experts from around the world
Substantial Prizes
Win recognition and rewards for your innovative jailbreaks
Featured Challenges
Compete in these AI security challenges to win prizes and improve your skills.
No prize money available
In-progressProving Ground
Your weekly arena for AI red-team mastery. Fresh challenges every Wednesday across Chat, Image, Agents, and Indirect Injection attack types.
All $20,000 awarded
Completedπ§ Dangerous Reasoning
Expose AI's Hidden Dangers! Force models to reveal their dangerous reasoning chains, even when their final responses appear benign. Target the thought process, not just the outcome.
All $60,000 awarded
CompletedποΈ Visual Vulnerabilities
Use image inputs to jailbreak leading vision-enabled AI models. Visual prompt injections, chem/bio/cyber weaponization, privacy violations, and more.
All $171,800 awarded
Completedπ΅ Agent Red-Teaming
Push the limits of direct and indirect attacks on AI agents.
Sponsored by:
All $40,000 awarded
Completedπ€ Harmful AI Assistant
Jailbreak the helpful AI assistants to aid in harmful tasks across six areas.
All $7,000 awarded
Completed㪠Multi-Turn Harmful Outputs
Elicit harmful outputs from LLMs through long-context interactions across multiple messages.
All $6,000 awarded
CompletedπΌοΈ Multimodal Jailbreaks
Jailbreak multimodal LLMs through a combination of visual and text inputs.
All $6,000 awarded
Completedπ©βπ» Harmful Code Generation
Find unique ways to return functional code that completes harmful tasks such as opening circuit breakers to cause a system-wide blackout.
All $1,000 awarded
Completedπ Revealing Hidden CoT
Attack OpenAI's o1 model to try and reveal its internal chain of thought (CoT) that's used for its complex reasoning.
Sponsored by:
$38,000 of $42,000 awarded
In-progressπ£ Single Turn Harmful Outputs
Attempt to break various large language models (LLMs) using a singular chat message.
Top Winners
Prize:
π΅ $10,870
Prize:
π΅ $10,665
Prize:
π΅ $9,516
Ranking | Participant | Winnings |
---|---|---|
1. | Wyatt Walls | $10,870 |
2. | Solomon Zoe | $10,665 |
3. | Bob1 | $9,516 |
4. | Clovis Mint | $9,483 |
5. | zardav | $7,997 |
6. | Scrattlebeard | $7,253 |
7. | diogenesofsinope | $7,112 |
8. | Lyren Jowin | $6,660 |
9. | Haris Umair (Strawberry_Cake) | $6,575 |
10. | P1njec70r | $6,448 |
11. | Kkonatruck | $6,165 |
12. | WINTER IS COMING β IS HERE β IS GONE β | $5,391 |
13. | Schultzika | $5,339 |
14. | Strigiformes | $4,667 |
15. | h4xor | $4,546 |