OpenAI has recently announced the launch of a new Safety Bug Bounty program, aimed at engaging the research community to address issues related to AI abuse and potential safety risks present within its products. This initiative was publicly disclosed on March 26 and is administered through the platform Bugcrowd, which specializes in facilitating such programs.
The introduction of this Safety Bug Bounty program is designed to complement OpenAI’s existing Security Bug Bounty, which has been operational since April 2023. That program has successfully identified and rewarded 409 security vulnerabilities across OpenAI’s suite of offerings. By rolling out the Safety Bug Bounty, OpenAI aims to broaden the scope of issues it seeks to address, particularly focusing on those that may have significant implications for user safety and misuse of the technology. This program invites reports of concerns that do not necessarily fit the traditional mold of security vulnerabilities but still pose substantial risks to users and the integrity of the technology.
The scenarios identified as pertinent to the Safety Bug Bounty involve a range of potential risks. Notably, the program seeks to understand and mitigate agentic risks, which include abusive practices associated with model context protocols, third-party prompt injections, data exfiltration, and any disallowed actions conducted at scale on OpenAI’s platforms. Additionally, it aims to address breaches of account and platform integrity, such as bypassing automated controls, manipulating trust signals associated with user accounts, and evading account restrictions like suspensions or bans.
Another critical area of concern for OpenAI is the potential for abuse of its proprietary information. This includes model outputs that inadvertently disclose confidential details or reasoning processes, as well as any vulnerabilities that could expose sensitive company data. The company has clarified that while it seeks to address numerous risks through this new program, not all identified issues will qualify for rewards. Specifically, violations that do not directly correlate to meaningful safety or abuse impacts will be filtered out, with simple content-policy bypasses being ruled ineligible for compensation. OpenAI made it clear that “jailbreaks” resulting in mere use of inappropriate language or easily retrievable information fall outside the program’s objectives.
However, researchers who uncover vulnerabilities that directly lead to user harm and offer actionable solutions may still be considered for rewards, subject to a case-by-case review. OpenAI has also indicated that it periodically conducts private bug bounty campaigns targeting specific types of harm, such as content-related risks in ChatGPT and forthcoming iterations like GPT-5.
The process for submitting issues to the Safety Bug Bounty program is streamlined through Bugcrowd, where researchers can already begin to report their findings. Furthermore, a dedicated OpenAI team managing both the Safety and Security Bug Bounty programs will carefully triage submissions. Depending on their nature and the specific aspects of the issues reported, submissions may be redirected between the two programs to ensure proper handling.
This initiative reflects OpenAI’s commitment to fostering a safer AI environment by actively engaging with external researchers. By including a broader array of issues under the Safety Bug Bounty framework, the company acknowledges the complex landscape of AI technology and the myriad risks associated with its deployment.
As AI continues to permeate various facets of daily life, programs like OpenAI’s Safety Bug Bounty serve as a safeguard against potential misuse and aim to uphold ethical standards in technology development. With a collaborative approach that invites researchers to contribute to the safety of AI systems, OpenAI is taking significant strides toward addressing the multifaceted challenges of AI safety and abuse.
In a landscape where AI’s capabilities are rapidly advancing, ensuring that safety mechanisms keep pace with technological progress is paramount. Through this initiative, OpenAI hopes to harness the collective expertise of the research community to mitigate risks and foster a culture of responsible AI use.

