jailbreak_llms
This study presents a comprehensive analysis of the largest in-the-wild jailbreak prompt collection from December 2022 to December 2023. Utilizing the JailbreakHub framework, it compiles data from platforms such as Reddit and Discord, focusing on the risks and behaviors linked to harmful language prompts. The dataset, containing over 15,000 prompts with 1,405 identified as jailbreak prompts, provides crucial insights into the vulnerabilities and possible safeguards of large language models.