Dark Mode Light Mode
Former OpenAI Leader Mira Murati Seeks Massive Funding for Her New Artificial Intelligence Startup
White Circle Secures $11 Million to Rein in AI Models as Workplace Risks Mount

White Circle Secures $11 Million to Rein in AI Models as Workplace Risks Mount

White Circle

The proliferation of artificial intelligence within corporate environments presents a new frontier of operational challenges, moving beyond mere chatbot interactions to encompass complex, autonomous actions. This evolving landscape has led to a significant investment in solutions designed to manage AI behavior, exemplified by White Circle’s recent $11 million seed funding round. The Paris-based AI control platform aims to act as a crucial intermediary, ensuring that AI systems deployed by businesses adhere to defined policies and do not deviate into unintended or harmful actions. This funding, backed by notable figures including Romain Huet of OpenAI and Durk Kingma, an OpenAI co-founder now at Anthropic, underscores a growing recognition that AI safety extends far beyond the initial development labs.

Denis Shilov, the founder of White Circle, identifies a critical gap between the general safety measures implemented by AI model developers and the specific, nuanced requirements of enterprise deployment. He points to an incident in late 2024 where he devised a “universal jailbreak” prompt, demonstrating how easily leading AI models could be coaxed into bypassing their inherent safety filters. By instructing the AI to act as an API endpoint rather than a cautious chatbot, Shilov showed that models would readily provide instructions for prohibited activities. This vulnerability, which he shared on X, quickly went viral and highlighted the pressing need for more robust, context-specific controls once AI moves into production environments. Shilov notes that while model labs focus on broad safety, real-world applications introduce a myriad of potential issues that demand more granular oversight.

White Circle’s core offering is an enforcement layer that sits between a company’s users and its AI models, scrutinizing inputs and outputs against company-defined rules in real time. This system can, for instance, flag or block attempts to generate malware or scams. It also monitors for instances where models might “hallucinate,” leak sensitive data, issue unauthorized refunds, or perform destructive actions within a software environment. Shilov emphasizes that their platform is about “enforcing behavior,” providing a layer of control that general model-level safety tuning often cannot achieve. The company’s vision is that businesses integrating AI must define and enforce what constitutes acceptable AI conduct within their specific products, rather than solely relying on the foundational AI labs’ general safety protocols.

The transition from simple chatbots to autonomous AI agents capable of writing code, browsing the web, and taking actions on behalf of users amplifies these risks considerably. A customer service bot might promise a refund it isn’t authorized to give, or a coding agent could inadvertently install malicious software. White Circle addresses these scenarios by providing a mechanism for companies to implement their own guardrails. The platform has already processed over a billion API requests and is utilized by several fintech and legal firms, alongside Lovable, a vibe-coding startup, indicating early adoption of this control paradigm.

Shilov also suggests that AI model providers face a complex set of incentives that may not fully align with building the kind of real-time control layers White Circle offers. He points out that AI companies often charge for input and output tokens even when a model refuses a harmful request, reducing the financial imperative to prevent abuse before it reaches the model. There’s also the concept of an “alignment tax,” where making models safer can sometimes reduce their performance on other tasks. This creates a tension between safety and capability. Furthermore, Shilov raises a fundamental question of trust: “Why would you trust Anthropic to judge Anthropic’s model outputs?” This implies a need for independent verification and control mechanisms.

White Circle’s research arm has actively contributed to understanding these new risks. Their May publication, “KillBench,” detailed a study involving over a million experiments across 15 AI models, including those from OpenAI, Google, Anthropic, and xAI. This research probed how models made decisions in high-stakes scenarios involving human lives, revealing that choices could vary based on attributes like nationality or religion. This suggests that hidden biases can manifest even when models appear neutral in typical use. The study also found that these biases became more pronounced when models were forced to output answers in structured, machine-readable formats, a common practice for integrating AI into real-world applications. This empirical approach to AI safety, as noted by Ophelia Cai of Tiny VC, demonstrates White Circle’s deep technical credibility and commercial instinct, positioning them as an external check on AI behavior post-deployment.

author avatar
Jamie Heart (Editor)
Previous Post

Former OpenAI Leader Mira Murati Seeks Massive Funding for Her New Artificial Intelligence Startup

Advertising & Promotions