Lewis Birch

Founding Engineer @ Mindgard
Adversarial ML Researcher · Lancaster University

Lewis Birch
2022 — Present

Mindgard

Founding Engineer

Building the AI security platform from day one — automated red teaming, guardrail evasion research, and adversarial attack tooling used to discover 70+ vulnerabilities in production AI systems from Microsoft, NVIDIA, Meta, and others.

2022 — Present

Lancaster University

PhD — Secure Machine Learning

Researching adversarial vulnerabilities in training data and ML model pipelines within the Experimental Distributed Systems lab.

2018 — 2022

Lancaster University

MSci Computer Science — First Class Honours

Four-year integrated master's with academic scholarship. Elected student representative.

LLM GuardrailsPrompt InjectionEvasion

Bypassing Prompt Injection and Jailbreak Detection in LLM Guardrails

Demonstrates two approaches for bypassing LLM guardrail systems via character injection and adversarial ML evasion techniques. Tested against six prominent systems including Microsoft Azure Prompt Shield and Meta Prompt Guard, achieving up to 100% evasion success in some cases.

Hackett, Birch, Trawicki, Suri, Garraghan · 2025 · LLMSEC Workshop · 25 citations
Read on arXiv
LLMModel ExtractionChatGPT

Model Leeching: An Extraction Attack Targeting LLMs

Introduces a novel extraction technique that distills task-specific knowledge from large language models into compact models. Achieved 73% exact match similarity with ChatGPT-3.5-Turbo at only $50 in API cost, with an 11% boost in adversarial attack transferability.

Birch, Hackett, Trawicki, Suri, Garraghan · 2023 · 34 citations
Read on arXiv
Deep LearningCompiler DefenseTensor Optimization

Compilation as a Defense: Enhancing DL Model Attack Robustness via Tensor Optimization

Proposes using model compilation and tensor optimization as a lightweight defense against side-channel attacks in adversarial ML — reducing attack effectiveness by up to 43% without costly model re-engineering.

Trawicki, Hackett, Birch, Suri, Garraghan · 2023 · 3 citations
Read on arXiv

Vulnerabilities discovered in production AI guardrail systems through security research at Mindgard.

Microsoft
Azure Prompt Shield — Guardrail Evasion
Guardrail Bypass
Microsoft
Azure AI Content Safety — Guardrail Evasion
Guardrail Bypass
NVIDIA
NemoGuard Jailbreak Detect — Guardrail Evasion
Guardrail Bypass
Meta
Prompt Guard — Guardrail Evasion
Guardrail Bypass
Protect AI
Jailbreak & Prompt Injection — Guardrail Evasion
Guardrail Bypass
Vijil
Prompt Injection — Guardrail Evasion
Guardrail Bypass

AI Under Attack: Six Key Adversarial Attacks and Their Consequences

mindgard.ai
Read →

Introduction to Adversarial Machine Learning

lewisbirch.dev
Read →