SafetyKit uses GPT-5 to detect multimodal fraud

3 minutes
OPENAI
SafetyKit uses GPT-5 to detect multimodal fraud

OpenAI published on September 9, 2025 a case study about SafetyKit, a startup that connects advanced models with agents designed to detect fraud and violations on marketplaces and payment platforms. Why does this matter now? Because it combines vision, text and transactional signals to make more accurate decisions at scale. (openai.com)

What SafetyKit announces

SafetyKit presents multimodal agents that review user content, images and financial operations to identify risk, fraud and prohibited content. The company says it reviews 100% of content and reaches over 95% accuracy in its internal evaluations, relying on models like GPT-5 and GPT-4.1, plus techniques such as RFT and a component they call CUA to automate complex tasks. (openai.com)

Sounds like marketing? A bit, yes. But it’s also a concrete example of how current models are used in product: it’s not just generating pretty text, it’s applying multimodal reasoning for regulated, high-impact decisions. (openai.com)

How they're applying it in practice

SafetyKit designs agents by risk type: one for scams, another for illegal products, another for legal disclosures, and so on. Each piece of content is routed to the agent that best fits that category and to the model most suitable for the task. For example, GPT-5 is used for multimodal reasoning when you need to understand visual and textual context; GPT-4.1 is used to follow policy instructions and handle high volumes. (openai.com)

An example that shows the difference: detecting a phone number embedded inside a product image or identifying a malicious QR. It’s not enough to search for keywords; the system must read the image, interpret intent and apply the correct rule. SafetyKit says their agents do exactly that. (openai.com)

"OpenAI gives us access to the most advanced multimodal reasoning models on the market. It lets us adapt quickly, launch new agents and process types of content other solutions can’t even parse."

David Graunke, founder and CEO of SafetyKit. (openai.com)

Scalability and results that matter

According to the note, SafetyKit’s platform went from processing 200 million tokens per day to 16 billion in six months, and expanded coverage to payment risks, fraud, child exploitation and money laundering, supporting clients with hundreds of millions of users. That reflects two things: growing demand for automated solutions and the technical capacity to integrate new model releases very quickly. (openai.com)

They also explain that each new model version, like o3 or GPT-5, is tested against the hardest cases and, if it improves results, deployed to production in days. That practice shortens the gap between research and product and can translate into more accurate detections in the field. (openai.com)

What you should think about if you work on a platform

  • Do you want to cover 100% of content without multiplying false positives? The combination of multimodal models plus specialized agents helps, but it depends on how routing rules and evaluations are designed.

  • Who audits the decisions? Internal metrics matter, but you should ask for external evaluations and transparency about real-world errors.

  • Privacy and compliance? Processing images and payment signals requires data controls and traceability, especially in regulated sectors.

In short, the technology opens real possibilities, but adoption requires governance, testing and monitoring. It’s not enough to add a powerful model and expect everything to improve by magic.

Reflective conclusion

This news shows a clear trend: risk and compliance teams are no longer just rules and blacklists. They now combine specialized agents with models that think across modes at once. The consequence? Greater coverage and, potentially, less burden on human moderators—provided platforms keep quality controls and accountability.

If you work at a marketplace or fintech, this isn’t a technical curiosity, it’s a product and risk decision that deserves testing and oversight.

If you want to read the original note, OpenAI has the announcement and SafetyKit maintains information about their service. SafetyKit and the OpenAI post offer more details. (openai.com)

Stay up to date!

Receive practical guides, fact-checks and AI analysis straight to your inbox, no technical jargon or fluff.

Your data is safe. Unsubscribing is easy at any time.