Skip to main content
PaperAI Ethics & Safetyv1.0

Weak-to-Strong Generalization: Eliciting Strong Capabilities With Weak Supervision

by OpenAI · free · Last verified 2026-03-17

Investigates whether strong models can be aligned using supervision from weaker models—an analogy to future superalignment challenges where humans supervise superintelligent AI. The paper demonstrates that naive weak-to-strong fine-tuning yields significant generalization beyond the weak supervisor and proposes auxiliary confidence loss and bootstrapping to close the performance gap.

https://arxiv.org/abs/2312.09390
B
BAbove Average
Adoption: B+Quality: A+Freshness: B+Citations: AEngagement: F

Specifications

License
Open Access
Pricing
free
Capabilities
scalable-oversight, weak-supervision, superalignment, alignment-research
Integrations
Use Cases
ai-safety-research, superalignment, research
API Available
No
Tags
alignment, superalignment, weak-supervision, scalable-oversight, generalization
Added
2026-03-17
Completeness
100%

Index Score

68
Adoption
74
Quality
92
Freshness
70
Citations
80
Engagement
0

Put AI to work for your business

Deploy this paper alongside autonomous AaaS agents that handle tasks end-to-end — no babysitting required.

Explore the full AI ecosystem on Agents as a Service