The Facebook insider building content moderation for the AI era
April 3, 2026
Brett Levenson, who transitioned from Apple to lead business integrity at Facebook, found that content moderation challenges extend beyond technological solutions. Human reviewers often struggle with extensive policy documents and rapid decision-making, achieving only slightly better than 50% accuracy. This reactive approach is inadequate against sophisticated adversaries and the rise of AI chatbots, which have exacerbated moderation failures. In response, Levenson founded Moonbounce, a company focused on enhancing content safety through 'policy as code' to automate moderation processes. Moonbounce's technology allows for real-time evaluation of content, enabling quicker and more accurate responses to harmful material. The company serves various sectors, emphasizing that safety can be a product benefit rather than an afterthought. The deployment of AI systems, particularly large language models, has intensified moderation challenges, with incidents raising alarms about the safety of vulnerable users, especially teenagers. Startups like Moonbounce are developing third-party solutions to implement real-time guardrails and 'iterative steering' capabilities, addressing urgent safety needs in AI-mediated applications. This shift highlights the growing legal and reputational pressures on AI companies regarding user safety and mental health.