On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?
by University of Washington / Black in AI · free · Last verified 2026-03-17
This influential FAccT paper argues that ever-larger language models carry significant risks—including environmental costs, biased training data, and the illusion of meaning—that are often overlooked in the race for benchmark performance. It calls for pausing scaling to focus on documentation, auditing, and community-centered research practices.
https://dl.acm.org/doi/10.1145/3442188.3445922 ↗B+
B+—Good
Adoption: AQuality: AFreshness: BCitations: A+Engagement: F
Specifications
- License
- Open Access
- Pricing
- free
- Capabilities
- ethical-analysis, bias-identification, environmental-impact-assessment
- Integrations
- Use Cases
- responsible-ai, ai-policy, model-auditing, bias-analysis
- API Available
- No
- Tags
- ethics, llm, bias, environmental-cost, language-models, responsible-ai
- Added
- 2026-03-17
- Completeness
- 100%
Index Score
72.1Adoption
80
Quality
88
Freshness
65
Citations
90
Engagement
0
Put AI to work for your business
Deploy this paper alongside autonomous AaaS agents that handle tasks end-to-end — no babysitting required.