Model Comparison Script
by AaaS · open-source · Last verified 2026-03-01
Side-by-side model comparison script that runs identical prompts through multiple LLM APIs and presents results in a structured format. Measures response quality, latency, token usage, and cost per query with automated scoring via LLM judges.
https://aaas.blog/script/model-comparison-script ↗C+
C+—Average
Adoption: BQuality: B+Freshness: B+Citations: C+Engagement: F
Specifications
- License
- MIT
- Pricing
- open-source
- Capabilities
- multi-model-inference, latency-measurement, cost-calculation, automated-scoring, report-generation
- Integrations
- openai, anthropic, pandas, aiohttp
- Use Cases
- model-selection, vendor-comparison, cost-optimization, quality-assessment
- API Available
- No
- Language
- python
- Dependencies
- openai, anthropic, pandas, aiohttp, rich
- Environment
- Python 3.11+
- Est. Runtime
- 5-20 minutes depending on prompt count and models
- Tags
- script, automation, comparison, evaluation, analysis
- Added
- 2026-03-17
- Completeness
- 100%
Index Score
53.3Adoption
64
Quality
76
Freshness
78
Citations
50
Engagement
0