What we are building

Open-source tools and research projects focused on AI security and safety. Everything we build starts from a real vulnerability or risk we found.

Adversarial RL Environment

Active

Reinforcement learning finally works for security. All you need is the right feedback loop. Our RL environments transform pretrained models into offensive and defensive security agents. Robust, packed with realistic attack scenarios, and designed with hundreds of programmatically verifiable security challenges at the edge of frontier capabilities.

AI Safety Research

Research

Research into how frontier AI systems fail - from reward hacking and hallucination cascades to alignment drift in autonomous pipelines. We build benchmarks and evaluation frameworks that map the boundaries of safe behavior across model families and real-world deployments.

Agent Security Research

Research

Threat modeling for autonomous AI agents that make decisions, call tools, and chain actions without human oversight. We study prompt injection, tool poisoning, memory manipulation, and privilege escalation - and publish the attack taxonomies and defense frameworks the industry relies on.

AI for Cybersecurity

Active

Applying AI to scale offensive security. We use language models and automated analysis to discover vulnerabilities in software - including two CVEs in Apple's software.

Interested in what we are building?

Schedule a call with the founders to discuss collaboration, research partnerships, or security audits.

Schedule a call →