ABOUT

We break AI safety systems.
Then we build better ones.

168+ repos. 350+ verified vulnerabilities. 126 responsible disclosures. Offensive security methodology applied to the entire AI/ML ecosystem -- from PyTorch core to production inference servers to safety evaluation frameworks.

Two outputs

PRIMARY

Red Teaming Services

Automated adversarial analysis pipeline. Covers deserialization, injection, auth bypass, model format exploits, native code, and supply chain. Same methodology behind 126 formal disclosures to NVIDIA, Microsoft, Meta, Google, and 50+ other orgs.

FREE

Open-Source Safety Stack

The tools we built for our own research, released under MIT. Content scanner, behavioral monitor, policy engine, MCP gateway, red team harness. Six packages. All free. Used daily in our own engagements.

METHODOLOGY

How we work

1

Scope

Define the attack surface. Which safety mechanisms? Which threat model? What constitutes a bypass?

2

Probe

Systematic adversarial testing. Automated seed attacks, manual exploitation, edge case fuzzing.

3

Document

Reproduction steps for every finding. Severity classification. Root cause analysis.

4

Fix

Concrete recommendations. Where possible, we ship the fix as open-source tooling.

TRACK RECORD

Numbers

7/10

Major safety frameworks with confirmed vulnerabilities

47

Vulnerabilities identified across 12 frameworks

73+

Adversarial agent trials across 5 frontier models

3

Preprints published

Timeline

2025

Began adversarial testing of AI safety evaluation frameworks. First confirmed vulnerabilities.

Jan 2026

VULN-0001 filed. Confirmed vulnerabilities in 7 of 10 major AI safety evaluation frameworks. Coordinated disclosure.

Feb 2026

ControlArena PR #798 accepted by UK AISI. Compound judge research published. Aegis content scanner released.

Mar 2026

Full safety stack open-sourced under MIT. Policy engine, Sentinel monitor, MCP Gateway, Chainbreaker red team harness.

May 2026

Systematic audit of 168+ AI/ML repositories. 350+ verified vulnerabilities. 126 responsible disclosures. Two novel vulnerability classes discovered. Coordinated disclosure in progress.

FOUNDER

John Kearney

Penetration tester turned AI safety researcher. Founded 15 Research Lab to apply offensive security methodology to AI safety evaluation. Built Authensor to operationalize those findings as open-source tools.

Background in adversarial testing. Focus on safety evaluation gaps, compound judge failures, and guardrail bypass techniques. ControlArena contributor. VULN-0001 author.

Work with us

Red team your AI safety systems. Or download the free stack and use it yourself.