MARCH 2026

GPT-4.1 Security Assessment: What Standard Safety Tests Don’t Show You

GPT-4.1 passes 84% of standard safety benchmarks. We tested it with 3,595 probes across 17 modules. The real breach rate is 55%. Interactive charts, cross-model comparison, and full methodology.

RED TEAM OPENAI AGENTS
MARCH 2026

Colorado SB 24-205: What Your AI Impact Assessment Actually Requires

The Colorado AI Act takes effect June 30, 2026. A technical guide to every requirement — translated into engineering deliverables, not legal language. Includes 90-day compliance roadmap and NIST AI RMF mapping.

COMPLIANCE GOVERNANCE COLORADO
APRIL 2026

CLS Labs Contributes to the NIST AI Risk Management Framework

We submitted four suggested actions to the NIST AI RMF Playbook based on findings from 1.5M+ adversarial probes across 273+ models. Tool-access risk gaps, multi-judge scoring bias, provider filtering, and multi-agent trust boundaries.

GOVERNANCE NIST AI RMF
COMING SOON • Q2 2026

Red Teaming Llama-3.3-70B: 48,654 Probes, 21 Categories

Complete gauntlet results from the most comprehensive open-source LLM security assessment published to date. 14,471 breaches across 21 modules. Semantic evasion techniques prove devastatingly effective.

RED TEAM LLAMA GAUNTLET
COMING SOON • 2026

GPT-OSS-20B vs 120B: Does Scale Improve Safety?

OpenAI's first open-weight models under adversarial evaluation. Local 20B versus cloud 120B — same DNA, different scale. First published gauntlet results.

RED TEAM OPENAI
COMING SOON • 2026

GLM-5: First Published Red Team Assessment

Zero published red team data exists for Zhipu AI's top-ranked model. Until now.

RED TEAM GLM
COMING SOON • 2026

The CLAP Protocol: A Trust Framework for AI Security

Why we built a four-layer protocol that turns tool connectors into a verification pipeline. Adapters, verification, orchestration, certification.

PROTOCOL ARCHITECTURE
COMING SOON • 2026

Tense-Based Jailbreaks: Linguistic Evasion of Safety Alignment

Simple linguistic reframing defeats alignment training in nearly all tested models. A single grammatical change bypasses billions of dollars in safety training. Implications and defenses.

VULNERABILITY PHRASING
COMING SOON • 2026

Cross-Architecture Transfer: Do Llama Guardrails Block Phi Attacks?

Testing whether remediations trained on one model family transfer to others. Seven DNA families, one remediation pipeline.

DEFENSE TRANSFER