Skip to content
SS

Sander Schulhof

AI Security Researcher and CEO of HackAPrompt

Sander Schulhoff is a leading AI researcher specializing in AI security, prompt injection, and red teaming. He authored the first comprehensive guide on prompt engineering and organized the first prompt injection competition, influencing AI labs and Fortune 500 companies. His work highlights the vulnerabilities in current AI guardrails and emphasizes the need for integrating classical cybersecurity expertise with AI knowledge.

Episodes (1)

Insights (10)

Camel Blocks Prompt Injection Through Permission Control

strategic thinking

Camel grants agents only the minimal read/write permissions inferred from the user’s request, blocking malicious actions introduced via prompt injection.

1h 4m

Education is Key Defense Against AI Attacks

leadership perspectives

Sander emphasises raising team awareness of prompt injection and hiring AI red-teaming experts, pointing listeners to a quarterly Maven course.

1h 9m

Three Security Tiers For AI Systems

strategic thinking

Sander lays out three security tiers—read-only chatbot, verified read-only with classical security, and agentic systems requiring extra defences against prompt injection.

59m

Remotely.io Twitter Bot Hijacked to Make Threats

case studies lessons

Remotely.io’s Twitter bot was prompt-injected to threaten the president, tarnishing the brand and leading to shutdown.

11m

First AI Red-Teaming Competition Created Benchmark Dataset

case studies lessons

Sander organised the first generative AI red-teaming event, creating a public prompt-injection dataset now used by frontier labs to benchmark security.

6m

AI Legislation Drives Compliance Industry

growth scaling tactics

With expanding AI legislation, firms like Trustable prosper by keeping companies updated on compliance and governance requirements.

1h 18m

AI Security Researcher on Team Validates Model Capabilities

leadership perspectives

Place an AI security researcher on the team to validate model abilities and cut through misinformation.

48m

Training AI Models Early for Adversarial Robustness

strategic thinking

Sander suggests injecting adversarial training early in the training stack—when the model is a “very small baby”—to raise baseline robustness.

1h 12m

Adaptive Evaluation for Adversarial Robustness

strategic thinking

Sander explains the best way to gauge adversarial robustness is an adaptive evaluation where your defence faces an attacker that learns and improves over time.

31m

Vegas Cybertruck Bombing Planner Used ChatGPT

case studies lessons

A Vegas bombing plotter reportedly jailbroke ChatGPT for bomb-building instructions, showing lone-user exploitation risk.

14m