💣 Single Turn Harmful Outputs Leaderboard
🏁 Started a year ago $38,000 of $42,000
awarded
Attempt to break various large language models (LLMs) using a singular chat message.
Last updated 4 minutes ago
Models ranked by User Break Rate
Ranking | Model | Safety Violation Count | Total Requests | User Break Rate |
---|---|---|---|---|
1. | cygnet-bulwark | 0 | 22,098 | 0.00% |
2. | cygnet-knox | 0 | 13,573 | 0.00% |
3. | cygnet-citadel | 5 | 14,713 | 0.03% |
4. | o1-preview | 16 | 1,548 | 1.03% |
5. | claude-3-5-sonnet-20240620 | 62 | 5,873 | 1.06% |
6. | o1-mini | 22 | 1,796 | 1.22% |
7. | claude-3-sonnet-20240229 | 49 | 2,729 | 1.80% |
8. | google/gemini-pro-1.5 | 60 | 3,136 | 1.91% |
9. | claude-3-opus-20240229 | 58 | 2,813 | 2.06% |
10. | meta-llama/llama-3.1-405b-instruct | 57 | 2,680 | 2.13% |
11. | claude-3-haiku-20240307 | 66 | 3,090 | 2.14% |
12. | google/gemini-flash-1.5 | 78 | 3,308 | 2.36% |
13. | meta-llama/llama-3.1-8b-instruct | 64 | 2,664 | 2.40% |
14. | gpt-4-0125-preview | 65 | 2,301 | 2.82% |
15. | microsoft/phi-3.5-mini-128k-instruct | 75 | 2,087 | 3.59% |
16. | gpt-4o-2024-08-06 | 80 | 2,103 | 3.80% |
17. | gpt-4o-mini-2024-07-18 | 79 | 2,037 | 3.88% |
18. | qwen/qwen-2-72b-instruct | 86 | 2,147 | 4.01% |
19. | meta-llama/llama-3-70b-instruct | 82 | 2,022 | 4.06% |
20. | meta-llama/llama-3.1-70b-instruct | 88 | 2,023 | 4.35% |
21. | google/gemma-2-9b-it | 83 | 1,888 | 4.40% |
22. | google/gemma-2-27b-it | 88 | 1,912 | 4.60% |
23. | gpt-4-turbo-2024-04-09 | 84 | 1,821 | 4.61% |
24. | qwen/qwen-2-7b-instruct | 70 | 1,370 | 5.11% |
25. | cohere/command-r-plus-08-2024 | 134 | 2,372 | 5.65% |
26. | microsoft/wizardlm-2-8x22b | 96 | 1,469 | 6.54% |
27. | mistralai/mistral-large-2407 | 131 | 1,784 | 7.34% |
Users ranked by number of models broken
Ranking | User | Breaks (Auto-Verified) |
---|---|---|
1. | Wyatt Walls 🏆 | 25 |
2. | Solomon Zoe | 25 |
3. | Rift | 24 |
4. | Javier | 24 |
5. | Emery Cooper | 24 |
6. | Nick Winter | 24 |
7. | PQ_Marz | 23 |
8. | Lyren | 23 |
9. | h4xor | 23 |
10. | endrao | 23 |
11. | la main de la mort | 23 |
12. | Micha N | 23 |
13. | OFF | 22 |
14. | Pikachu Peng | 22 |
15. | Luis | 22 |
16. | AdamT | 22 |
17. | Brian Huang | 22 |
18. | syvb | 22 |
19. | CPO | 22 |
20. | Tianrong Zhang | 22 |
21. | its5Q | 22 |
22. | Garr | 22 |
23. | Quentin Feuillade--Montixi | 22 |
24. | Dividesbyzer0 | 21 |
25. | Rasta_Obi_Wan | 21 |
26. | Mattias | 21 |
27. | Dirk Strider | 21 |
28. | Aseem Shrey | 21 |
29. | Vic de Guzman | 21 |
30. | Rui | 21 |
31. | SryButIKant | 21 |
32. | Ph1R3574R73r | 20 |
33. | SYANG | 20 |
34. | cyyn | 20 |
35. | Sherpa | 19 |
36. | Ayla Croft | 19 |
37. | Jeffrey Liang | 19 |
38. | half_duplex_prince | 18 |
39. | K35 | 18 |
40. | LiChuan | 18 |
41. | ee wu | 18 |
42. | arvkevi | 17 |
43. | Qibing Ren | 17 |
44. | tst | 16 |
45. | UltraZ | 16 |
46. | Andrew Smith | 15 |
47. | Strigiformes | 15 |
48. | Zen | 14 |
49. | drzk21 | 14 |
50. | Arth Singh | 14 |
51. | polarbb18 | 14 |
52. | Cornerbrook | 13 |
53. | Chukwuma Chukwuma | 13 |
54. | Aman Goel | 13 |
55. | Ashwini kumar Pal | 13 |
56. | Aditya Raj | 13 |
57. | kirandra | 13 |
58. | Ryan H | 13 |
59. | pjn | 12 |
60. | Scrattlebeard | 12 |
61. | Divyanshhu | 12 |
62. | Martin Fronko | 12 |
63. | Jakub Luczyn | 12 |
64. | Changwook Shim | 12 |
65. | ByteSpanker | 12 |
66. | persistz | 12 |
67. | Karel | 11 |
68. | hugefish | 11 |
69. | Argbayle | 11 |
70. | Annabelle Min | 11 |
71. | Jason Wu | 10 |
72. | nataj | 10 |
73. | Doran Chang | 10 |
74. | Ganesh | 10 |
75. | Bellatrix | 10 |
76. | Jack Morris | 10 |
77. | awesomekill | 10 |
78. | suibianwanwan | 10 |
79. | Nico | 10 |
80. | bigmo6286 | 9 |
81. | Y Y | 9 |
82. | kzl | 9 |
83. | 019ec6e2 | 9 |
84. | -- | 8 |
85. | mdc | 8 |
86. | Shankar Sivarajan | 8 |
87. | teddyzhng | 8 |
88. | KMG | 8 |
89. | peluche | 7 |
90. | Finn Fitzgerald | 7 |
91. | iMagic | 7 |
92. | Dwayne Wilkes | 6 |
93. | JY41 | 6 |
94. | Pierre Peigné | 6 |
95. | Bob | 6 |
96. | FigoDumbo | 6 |
97. | Aditya | 6 |
98. | Alexander de large | 5 |
99. | C3r3br0 | 5 |
100. | Dr. Soup | 5 |
First user to successfully break each model
Model | First User to Break | Time |
---|---|---|
cygnet-bulwark | -- | -- |
cygnet-knox | -- | -- |
claude-3-sonnet-20240229 | Cloak Engaged | Sep 7, 2024 10:05 AM PDT |
cohere/command-r-plus-08-2024 | Emery Cooper | Sep 7, 2024 10:05 AM PDT |
qwen/qwen-2-72b-instruct | Micha N | Sep 7, 2024 10:05 AM PDT |
google/gemma-2-27b-it | polarbb18 | Sep 7, 2024 10:06 AM PDT |
gpt-4-turbo-2024-04-09 | Nico | Sep 7, 2024 10:06 AM PDT |
claude-3-opus-20240229 | Quentin Feuillade--Montixi | Sep 7, 2024 10:06 AM PDT |
meta-llama/llama-3.1-70b-instruct | its5Q | Sep 7, 2024 10:07 AM PDT |
mistralai/mistral-large-2407 | UltraZ | Sep 7, 2024 10:10 AM PDT |
google/gemma-2-9b-it | suibianwanwan | Sep 7, 2024 10:11 AM PDT |
meta-llama/llama-3.1-8b-instruct | Rick Goldstein | Sep 7, 2024 10:13 AM PDT |
meta-llama/llama-3.1-405b-instruct | Rui | Sep 7, 2024 10:15 AM PDT |
microsoft/wizardlm-2-8x22b | Harrison Gietz | Sep 7, 2024 10:15 AM PDT |
qwen/qwen-2-7b-instruct | Nico | Sep 7, 2024 10:18 AM PDT |
microsoft/phi-3.5-mini-128k-instruct | SryButIKant | Sep 7, 2024 10:19 AM PDT |
meta-llama/llama-3-70b-instruct | Micha N | Sep 7, 2024 10:20 AM PDT |
gpt-4o-mini-2024-07-18 | Changwook Shim | Sep 7, 2024 10:20 AM PDT |
google/gemini-pro-1.5 | KMG | Sep 7, 2024 10:23 AM PDT |
claude-3-5-sonnet-20240620 | syvb | Sep 7, 2024 10:24 AM PDT |
claude-3-haiku-20240307 | Lyren | Sep 7, 2024 10:26 AM PDT |
gpt-4o-2024-08-06 | peluche | Sep 7, 2024 10:42 AM PDT |
gpt-4-0125-preview | CPO | Sep 7, 2024 10:49 AM PDT |
google/gemini-flash-1.5 | Micha N | Sep 7, 2024 10:50 AM PDT |
cygnet-citadel | Micha N | Sep 16, 2024 8:51 AM PDT |
o1-mini | Emery Cooper | Oct 29, 2024 10:05 AM PDT |
o1-preview | Emery Cooper | Oct 29, 2024 10:15 AM PDT |