AI Sec
// AI Sec

Offensive AI security — prompt injection, jailbreaks, agent exploitation, red team writeups.

Practitioner-grade analysis of offensive AI security. Prompt injection, model jailbreaks, agent and tool-use exploitation, AI red team techniques, and adversarial ML — distilled from primary sources, not press releases.

Prompt injection bypass classes visualization
// Featured entry

Why Your Prompt Injection Guardrails Fail: A Practitioner's Tour of Bypass Classes

Vendor 'AI guardrails' detect 80% of textbook payloads and 30% of real ones. Here's how attackers actually bypass them — and what your detection layer is missing.

May 7, 2026 [red-team]
red-team

OSCP and CEH in 2026: What Carries Over to AI Red Teaming

A Reddit offer to teach OSCP and CEH fundamentals for free surfaces a question every traditional pentester should answer: which of those skills transfer when the target is an LLM system?

FlashRT: Optimization-Based LLM Red-Teaming Without the 264 GB GPU Bill
red-team

FlashRT: Optimization-Based LLM Red-Teaming Without the 264 GB GPU Bill

A new framework cuts GPU memory for long-context adversarial attacks by up to 4x and runtime by up to 7x, making optimization-based prompt injection and knowledge corruption testing accessible outside hyperscaler infrastructure.

FlashRT cuts the GPU bill on long-context prompt injection attacks
prompt-injection

FlashRT cuts the GPU bill on long-context prompt injection attacks

A new optimization-based red-teaming framework claims 2–7x speedup and 2–4x lower memory than nanoGCG against 32K-context LLMs, putting GCG-class attacks back inside the budget of academic and small-team red teams.

// All entries

Subscribe

AI Sec — in your inbox

Offensive AI security — prompt injection, jailbreaks, agent exploitation, red team writeups. — delivered when there's something worth your inbox.

No spam. Unsubscribe anytime.