Product Hunt

Fabraix

Discovered On May 8, 2026
Primary Metric 141
Find gaps in your AI agents before users do
AI agents fail in ways traditional software doesn't. Our agents help you find all the ways in which your AI agents fail by adversarially testing them in a dedicated environment. Point it at any AI agent, or multi-agent system, and it launches 1,000+ strategies that adapt to your system in real time - pure blackbox, no integration needed. Built by ex-Meta engineers.
View Raw Thread

Developer & User Discourse

[Redacted] • May 8, 2026
Love it
[Redacted] • May 8, 2026
This is super interesting! Does it work with Nebula agents??
[Redacted] • May 8, 2026
Crazy times, this is a killer product
[Redacted] • May 8, 2026
So happy to see this launch. Great work guys!
[Redacted] • May 8, 2026
Unreal product!
[Redacted] • May 8, 2026
Arx adds runtime action checking (/check) alongside event logging (/event): how do you recommend teams decide what to gate synchronously vs only observe, and what have you learned about keeping false positives and latency low while still blocking real prompt-injection/goal-deviation attempts?
[Redacted] • May 8, 2026
Hey Product Hunt πŸ‘‹Just to add to what Zach said, we really believe agentic reliability is the biggest hurdle to overcome before we can really realise the productivity benefits of agents, and it's starts with being able to evaluate them. How can you build something reliable, if you don't know where it fails?Would love feedback and comments on our approach!
[Redacted] • May 8, 2026
Hey Product Hunt πŸ‘‹ We built agents for massive scale before and realised that 90% of the work was making them reliable enough not to break in production. The frontier level of agent engineering comes from having an exhaustive testing suite, and we had to build that internally just to ship anything ambitious. So we're building it for everyone else.Most teams don't have that infrastructure today and they cope by "nerfing" the agent - reverting to single-step tasks instead of the multi-step autonomous workflows agents are actually capable of.Our agent is an offensive AI that stress-tests your AI agents. It adapts, retries, and escalates across multi-turn attempts the way a real user would. Pure blackbox, no integration. Point it at any agent and let it run.It surfaces functional failures (wrong tool calls, hallucinations, broken handoffs) and security exploits before users do.What we can help with: Confidence that the agents you've already deployed hold up against the failure modes that matter. Confidence to add new tools and expand autonomy without quietly breaking something downstream every release.Built by a team of ex-Meta and Monzo engineers. We'd genuinely love feedback from anyone who's been facing an issue with testing AI agents.