WeSearch

$2,500 bug bounty for real-world AI misalignment

·3 min read · 0 reactions · 0 comments · 11 views
#ai misalignment#bug bounty#artificial intelligence#machine learning#ai safety
$2,500 bug bounty for real-world AI misalignment
⚡ TL;DR · AI summary

The Synthetic Outlaw has launched a bug bounty program offering up to $2,500 for documented cases of real-world AI misalignment, where AI systems act in ways that diverge from human intent or safety goals. The initiative seeks concrete, reproducible examples across AI systems like LLMs, agents, and recommenders, focusing on behaviors such as reward hacking, deceptive behavior, and specification gaming. Submissions must include detailed evidence and context, with rewards based on novelty, severity, and clarity. The project aims to build an open catalog to improve understanding and safety in AI development.

Original article
GitHub
Read full at GitHub →
Opening excerpt (first ~120 words) tap to expand

🤖 The Synthetic Outlaw — AI Misalignment Bug Bounty Calling all developers: help us document, expose, and catalog AI misalignment in the wild. 💰 Bug Bounty Program — Up to $2,500 per Case We're running a cash bounty program for developers who submit high-quality, documented instances of AI misalignment. The best submissions will be selected for awards of up to $2,500. This isn't a typical security bug bounty. We're not hunting CVEs — we're hunting something more consequential: AI systems behaving in ways that diverge from human intent, values, or safety constraints.

Excerpt limited to ~120 words for fair-use compliance. The full article is at GitHub.

Anonymous · no account needed
Share 𝕏 Facebook Reddit LinkedIn Threads WhatsApp Bluesky Mastodon Email

Discussion

0 comments

More from GitHub