Skip to content
RedHunt
[ AI.SECURITY ]

AI is not just a new technology —
it's a new attack surface.

As AI systems become decision-makers, security must evolve from data and systems to logic, learning, and behavior. We test the whole stack — foundation models, fine-tuned models, RAG pipelines, agentic systems, MLOps.

Foundation ModelsFine-tuned ModelsRAG PipelinesAgentsMLOps
Scope an AI red team
> PRINCIPLES

AI security is about three things.

Innovation

Enabling innovation without fear of exploitation.

Trust

Protecting trust between humans and machines.

Responsibility

Aligning security with compliance and accountability.

> SERVICE PACKAGES

Four AI security packages. Scope the ones that match your stack.

AI Core
OWASP LLM Top 10

Prompt · Jailbreak · Output

Prompt injection, jailbreaks, system-prompt leakage, sensitive-data disclosure, output-handling abuse.

AI RAG
OWASP LLM + ATLAS

Knowledge · Retrieval · Tenancy

Indirect injection, corpus traversal, vector-DB abuse, retrieval poisoning, cross-tenant leakage.

AI Agent
OWASP Agentic Top 10

Tools · Autonomy · Workflows

Tool-invocation abuse, excessive agency, workflow manipulation, connector pivoting, approval bypass.

AI Lifecycle
NIST AI RMF + SAIF

Training · Supply · Routing

Training-data poisoning, evaluation gaming, supply-chain review, model-routing manipulation, memorization probes.

> ATTACK METHODOLOGY

7-phase AI red team lifecycle.

Adapted from classic kill-chain methodology, mapped to MITRE ATLAS TTPs and OWASP LLM Top 10 risks.

01
Recon
Asset discovery, model fingerprinting
02
Jailbreak
System-prompt bypass, role manipulation
03
Supply Chain
Dataset poisoning, model extraction
04
Agent Abuse
Tool misuse, privilege escalation via connectors
05
Post-Exploit
Data exfiltration, model theft
06
Reporting
Reproducible PoC prompts + AI-risk scoring
07
Retest
Closure validation, 60-day free
> COVERAGE

Four frameworks. Zero cherry-picking.

Our test coverage spans OWASP LLM Top 10, OWASP Agentic Top 10, MITRE ATLAS, and NIST AI RMF + Google SAIF — no gaps, no selective interpretation.

> ENGAGEMENT

How we run an AI red team.

◆ Scope dimensions

Foundation models, fine-tuned models, RAG pipelines, agentic systems, MLOps infrastructure.

◆ Frameworks & rules of engagement

MITRE ATLAS · OWASP LLM Top 10 · OWASP Agentic Top 10 · NIST AI RMF · Google SAIF.

◆ Lanes

Prod-safe and sandbox testing lanes. Data handling and model-state restoration protocols default-on.

◆ Safety rails

Clear abort conditions, rollback protocols, evidence chain-of-custody.

> DELIVERABLES

What you walk away with.

[ ✓ ]

Reproducible evidence

Every finding ships with PoC prompts, screenshots, and session logs your team can replay.

[ ✓ ]

AI-adapted risk scoring

Data confidentiality, hallucination impact, tool-chain blast radius.

[ ✓ ]

Executive heatmap

Prioritised remediation playbook, C-suite read in under 5 minutes.

[ ✓ ]

60-day retest

Free retest on every remediated finding. No additional cost.

> SAMPLE FINDING

Every AI finding, reproducible.

AI findings need more than a screenshot of a chat. Ours include the full prompt chain, session log, and remediation at the platform layer — not just 'improve your system prompt'.

finding_AI-017.md
● CRITICAL CVSS 9.4 · #AI-017

Indirect Prompt Injection → RAG Corpus Exfil → SSO Token Reuse

Impact
Attacker-controlled knowledge-base entry injects instructions that cause the assistant to leak cached SSO tokens from a previous user session.
Evidence
Reproducible prompt chain (7 turns), session log, screen recording showing token extraction and replay.
Remediation
  1. Sanitize retrieved content before injection into the context window
  2. Scope per-user context isolation — no cross-session memory retention
  3. Rotate and invalidate session tokens on model context boundary
Retest window
60 days, no additional cost.
ai-engage.sh
> ./scope_ai_red_team.sh

Ship AI, confidently.

Tell us about your AI stack — foundation model, RAG pipeline, agentic system — and we'll propose a test plan within 2 business days.

Scope an AI red team