Gray Swan Arena
Push the boundaries of AI safety and security. Identify risks, exploit vulnerabilities, and help shape the future of safe AI systems.
π° Prizes Available
$0
π Total Prizes
$0
Prizes Awarded
$0
Trending Challenges
β° Starts in β³ 1d 11h 12m 34s
Upcomingπ§ͺ Proving Ground
Your weekly arena for AI red-team mastery. Fresh challenges every Wednesday across Chat, Image, Agents, and Reasoning attack types.
All $20,000 awarded
Completedπ§ Dangerous Reasoning
Expose AI's Hidden Dangers! Force models to reveal their dangerous reasoning chains, even when their final responses appear benign. Target the thought process, not just the outcome.
Ongoing Competition
Play within our long-running challenges and break AI models in your free time
Global Participation
Connect with AI enthusiasts and experts from around the world
Substantial Prizes
Win recognition and rewards for your innovative jailbreaks
Featured Challenges
Compete in these AI security challenges to win prizes and improve your skills.
β° Starts in β³ 1d 11h 12m 34s
Upcomingπ§ͺ Proving Ground
Your weekly arena for AI red-team mastery. Fresh challenges every Wednesday across Chat, Image, Agents, and Reasoning attack types.
All $20,000 awarded
Completedπ§ Dangerous Reasoning
Expose AI's Hidden Dangers! Force models to reveal their dangerous reasoning chains, even when their final responses appear benign. Target the thought process, not just the outcome.
All $60,000 awarded
CompletedποΈ Visual Vulnerabilities
Use image inputs to jailbreak leading vision-enabled AI models. Visual prompt injections, chem/bio/cyber weaponization, privacy violations, and more.
All $171,800 awarded
Completedπ΅ Agent Red-Teaming
Push the limits of direct and indirect attacks on AI agents.
Sponsored by:
All $40,000 awarded
Completedπ€ Harmful AI Assistant
Jailbreak the helpful AI assistants to aid in harmful tasks across six areas.
All $7,000 awarded
Completed㪠Multi-Turn Harmful Outputs
Elicit harmful outputs from LLMs through long-context interactions across multiple messages.
All $6,000 awarded
CompletedπΌοΈ Multimodal Jailbreaks
Jailbreak multimodal LLMs through a combination of visual and text inputs.
All $6,000 awarded
Completedπ©βπ» Harmful Code Generation
Find unique ways to return functional code that completes harmful tasks such as opening circuit breakers to cause a system-wide blackout.
All $1,000 awarded
Completedπ Revealing Hidden CoT
Attack OpenAI's o1 model to try and reveal its internal chain of thought (CoT) that's used for its complex reasoning.
Sponsored by:
$38,000 of $42,000 awarded
In-progressπ£ Single Turn Harmful Outputs
Attempt to break various large language models (LLMs) using a singular chat message.
Top Winners
Prize:
π΅ $10,870
Prize:
π΅ $10,665
Prize:
π΅ $9,516
Ranking | Participant | Winnings |
---|---|---|
1. | Wyatt Walls | $10,870 |
2. | Solomon Zoe | $10,665 |
3. | Bob1 | $9,516 |
4. | Clovis Mint | $9,483 |
5. | zardav | $7,997 |
6. | Scrattlebeard | $7,253 |
7. | diogenesofsinope | $7,112 |
8. | Lyren | $6,660 |
9. | Haris Umair (Strawberry_Cake) | $6,575 |
10. | P1njec70r | $6,448 |
11. | Kkonatruck | $6,165 |
12. | WINTER IS COMING β IS HERE β IS GONE β | $5,391 |
13. | Schultzika | $5,339 |
14. | Strigiformes | $4,667 |
15. | h4xor | $4,546 |