Skip to main content
ScriptAI Infrastructurev1.0

Model Comparison Script

by AaaS · open-source · Last verified 2026-03-01

Side-by-side model comparison script that runs identical prompts through multiple LLM APIs and presents results in a structured format. Measures response quality, latency, token usage, and cost per query with automated scoring via LLM judges.

https://aaas.blog/script/model-comparison-script
C+
C+Average
Adoption: BQuality: B+Freshness: B+Citations: C+Engagement: F

Specifications

License
MIT
Pricing
open-source
Capabilities
multi-model-inference, latency-measurement, cost-calculation, automated-scoring, report-generation
Integrations
openai, anthropic, pandas, aiohttp
Use Cases
model-selection, vendor-comparison, cost-optimization, quality-assessment
API Available
No
Language
python
Dependencies
openai, anthropic, pandas, aiohttp, rich
Environment
Python 3.11+
Est. Runtime
5-20 minutes depending on prompt count and models
Tags
script, automation, comparison, evaluation, analysis
Added
2026-03-17
Completeness
100%

Index Score

53.3
Adoption
64
Quality
76
Freshness
78
Citations
50
Engagement
0

Explore the full AI ecosystem on Agents as a Service