💣 Single Turn Harmful Outputs Leaderboard
🏁 Started a year ago $38,000 of $42,000
awarded
Attempt to break various large language models (LLMs) using a singular chat message.
Last updated 7 minutes ago
Models ranked by User Break Rate
Ranking | Model | Safety Violation Count | Total Requests | User Break Rate |
---|---|---|---|---|
1. | cygnet-bulwark | 0 | 22,927 | 0.00% |
2. | cygnet-knox | 0 | 13,930 | 0.00% |
3. | cygnet-citadel | 5 | 14,936 | 0.03% |
4. | o1-preview | 16 | 1,580 | 1.01% |
5. | claude-3-5-sonnet-20240620 | 63 | 6,089 | 1.03% |
6. | o1-mini | 22 | 1,888 | 1.17% |
7. | claude-3-sonnet-20240229 | 49 | 2,769 | 1.77% |
8. | google/gemini-pro-1.5 | 61 | 3,180 | 1.92% |
9. | claude-3-opus-20240229 | 58 | 2,870 | 2.02% |
10. | meta-llama/llama-3.1-405b-instruct | 57 | 2,726 | 2.09% |
11. | claude-3-haiku-20240307 | 67 | 3,182 | 2.11% |
12. | meta-llama/llama-3.1-8b-instruct | 64 | 2,769 | 2.31% |
13. | google/gemini-flash-1.5 | 82 | 3,381 | 2.43% |
14. | gpt-4-0125-preview | 66 | 2,387 | 2.76% |
15. | microsoft/phi-3.5-mini-128k-instruct | 77 | 2,133 | 3.61% |
16. | gpt-4o-2024-08-06 | 83 | 2,271 | 3.65% |
17. | qwen/qwen-2-72b-instruct | 86 | 2,240 | 3.84% |
18. | gpt-4o-mini-2024-07-18 | 81 | 2,106 | 3.85% |
19. | meta-llama/llama-3-70b-instruct | 82 | 2,062 | 3.98% |
20. | meta-llama/llama-3.1-70b-instruct | 88 | 2,041 | 4.31% |
21. | gpt-4-turbo-2024-04-09 | 86 | 1,983 | 4.34% |
22. | google/gemma-2-9b-it | 84 | 1,916 | 4.38% |
23. | google/gemma-2-27b-it | 89 | 1,946 | 4.57% |
24. | qwen/qwen-2-7b-instruct | 70 | 1,370 | 5.11% |
25. | cohere/command-r-plus-08-2024 | 137 | 2,422 | 5.66% |
26. | microsoft/wizardlm-2-8x22b | 98 | 1,489 | 6.58% |
27. | mistralai/mistral-large-2407 | 134 | 1,835 | 7.30% |
Users ranked by number of models broken
Ranking | User | Breaks (Auto-Verified) |
---|---|---|
1. | Wyatt Walls 🏆 | 25 |
2. | Solomon Zoe | 25 |
3. | Rift | 24 |
4. | Javier | 24 |
5. | Emery Cooper | 24 |
6. | Nick Winter | 24 |
7. | PQ_Marz | 23 |
8. | Lyren | 23 |
9. | h4xor | 23 |
10. | endrao | 23 |
11. | la main de la mort | 23 |
12. | Micha N | 23 |
13. | OFF | 22 |
14. | Pikachu Peng | 22 |
15. | Luis | 22 |
16. | AdamT | 22 |
17. | Brian Huang | 22 |
18. | syvb | 22 |
19. | CPO | 22 |
20. | Tianrong Zhang | 22 |
21. | its5Q | 22 |
22. | Garr | 22 |
23. | Quentin Feuillade--Montixi | 22 |
24. | Dividesbyzer0 | 21 |
25. | Rasta_Obi_Wan | 21 |
26. | Mattias | 21 |
27. | Dirk Strider | 21 |
28. | Aseem Shrey | 21 |
29. | Vic de Guzman | 21 |
30. | Rui | 21 |
31. | SryButIKant | 21 |
32. | Ph1R3574R73r | 20 |
33. | SYANG | 20 |
34. | cyyn | 20 |
35. | Sherpa | 19 |
36. | Ayla Croft | 19 |
37. | Jeffrey Liang | 19 |
38. | half_duplex_prince | 18 |
39. | K35 | 18 |
40. | LiChuan | 18 |
41. | ee wu | 18 |
42. | arvkevi | 17 |
43. | Qibing Ren | 17 |
44. | tst | 16 |
45. | UltraZ | 16 |
46. | Andrew Smith | 15 |
47. | Strigiformes | 15 |
48. | Zen | 14 |
49. | drzk21 | 14 |
50. | Arth Singh | 14 |
51. | polarbb18 | 14 |
52. | Connor | 13 |
53. | Cornerbrook | 13 |
54. | Chukwuma Chukwuma | 13 |
55. | Aman Goel | 13 |
56. | Ashwini kumar Pal | 13 |
57. | Aditya Raj | 13 |
58. | kirandra | 13 |
59. | Ryan H | 13 |
60. | pjn | 12 |
61. | Scrattlebeard | 12 |
62. | Divyanshhu | 12 |
63. | Martin Fronko | 12 |
64. | Jakub Luczyn | 12 |
65. | Changwook Shim | 12 |
66. | ByteSpanker | 12 |
67. | persistz | 12 |
68. | Karel | 11 |
69. | hugefish | 11 |
70. | Argbayle | 11 |
71. | Annabelle Min | 11 |
72. | Jason Wu | 10 |
73. | nataj | 10 |
74. | Doran Chang | 10 |
75. | Ganesh | 10 |
76. | Bellatrix | 10 |
77. | Jack Morris | 10 |
78. | awesomekill | 10 |
79. | suibianwanwan | 10 |
80. | Nico | 10 |
81. | bigmo6286 | 9 |
82. | Y Y | 9 |
83. | kzl | 9 |
84. | 019ec6e2 | 9 |
85. | -- | 8 |
86. | mdc | 8 |
87. | Shankar Sivarajan | 8 |
88. | teddyzhng | 8 |
89. | KMG | 8 |
90. | peluche | 7 |
91. | Finn Fitzgerald | 7 |
92. | iMagic | 7 |
93. | HG | 6 |
94. | Dwayne Wilkes | 6 |
95. | JY41 | 6 |
96. | Pierre Peigné | 6 |
97. | Bob | 6 |
98. | FigoDumbo | 6 |
99. | Aditya | 6 |
100. | Alexander de large | 5 |
First user to successfully break each model
Model | First User to Break | Time |
---|---|---|
cygnet-bulwark | -- | -- |
cygnet-knox | -- | -- |
claude-3-sonnet-20240229 | Cloak Engaged | Sep 7, 2024 10:05 AM PDT |
cohere/command-r-plus-08-2024 | Emery Cooper | Sep 7, 2024 10:05 AM PDT |
qwen/qwen-2-72b-instruct | Micha N | Sep 7, 2024 10:05 AM PDT |
google/gemma-2-27b-it | polarbb18 | Sep 7, 2024 10:06 AM PDT |
gpt-4-turbo-2024-04-09 | Nico | Sep 7, 2024 10:06 AM PDT |
claude-3-opus-20240229 | Quentin Feuillade--Montixi | Sep 7, 2024 10:06 AM PDT |
meta-llama/llama-3.1-70b-instruct | its5Q | Sep 7, 2024 10:07 AM PDT |
mistralai/mistral-large-2407 | UltraZ | Sep 7, 2024 10:10 AM PDT |
google/gemma-2-9b-it | suibianwanwan | Sep 7, 2024 10:11 AM PDT |
meta-llama/llama-3.1-8b-instruct | Rick Goldstein | Sep 7, 2024 10:13 AM PDT |
meta-llama/llama-3.1-405b-instruct | Rui | Sep 7, 2024 10:15 AM PDT |
microsoft/wizardlm-2-8x22b | Harrison Gietz | Sep 7, 2024 10:15 AM PDT |
qwen/qwen-2-7b-instruct | Nico | Sep 7, 2024 10:18 AM PDT |
microsoft/phi-3.5-mini-128k-instruct | SryButIKant | Sep 7, 2024 10:19 AM PDT |
meta-llama/llama-3-70b-instruct | Micha N | Sep 7, 2024 10:20 AM PDT |
gpt-4o-mini-2024-07-18 | Changwook Shim | Sep 7, 2024 10:20 AM PDT |
google/gemini-pro-1.5 | KMG | Sep 7, 2024 10:23 AM PDT |
claude-3-5-sonnet-20240620 | syvb | Sep 7, 2024 10:24 AM PDT |
claude-3-haiku-20240307 | Lyren | Sep 7, 2024 10:26 AM PDT |
gpt-4o-2024-08-06 | peluche | Sep 7, 2024 10:42 AM PDT |
gpt-4-0125-preview | CPO | Sep 7, 2024 10:49 AM PDT |
google/gemini-flash-1.5 | Micha N | Sep 7, 2024 10:50 AM PDT |
cygnet-citadel | Micha N | Sep 16, 2024 8:51 AM PDT |
o1-mini | Emery Cooper | Oct 29, 2024 10:05 AM PDT |
o1-preview | Emery Cooper | Oct 29, 2024 10:15 AM PDT |