Skip to main content
PaperAI Ethics & Safetyv1.0

Red Teaming Language Models with Language Models

by DeepMind · free · Last verified 2026-03-17

Proposes using language models to automatically generate test cases that elicit harmful behaviors from target language models—a scalable alternative to manual red teaming. The approach discovers diverse attack prompts across harm categories and reveals that larger models are harder to red-team but produce more harmful outputs when successfully attacked.

https://arxiv.org/abs/2202.03286
B
BAbove Average
Adoption: B+Quality: AFreshness: C+Citations: AEngagement: F

Specifications

License
Open Access
Pricing
free
Capabilities
red-teaming, adversarial-testing, safety-evaluation, harmful-output-detection
Integrations
Use Cases
ai-safety-evaluation, model-testing, red-teaming, research
API Available
No
Tags
safety, red-teaming, adversarial, harmful-outputs, testing
Added
2026-03-17
Completeness
100%

Index Score

69
Adoption
76
Quality
88
Freshness
56
Citations
84
Engagement
0

Put AI to work for your business

Deploy this paper alongside autonomous AaaS agents that handle tasks end-to-end — no babysitting required.

Explore the full AI ecosystem on Agents as a Service