Groq
by Groq · paid · Last verified 2026-03-17
Groq is a semiconductor company that developed the Language Processing Unit (LPU), a custom chip for ultra-fast AI inference. Their managed API provides some of the fastest publicly available LLM inference speeds, often exceeding 800 tokens/second, making it ideal for latency-sensitive applications.
https://groq.com ↗B
B—Above Average
Adoption: BQuality: AFreshness: ACitations: B+Engagement: F
Specifications
- License
- Proprietary
- Pricing
- paid
- Capabilities
- Ultra-fast LLM inference, Proprietary LPU hardware architecture, Managed inference API, OpenAI-compatible API endpoint, High tokens-per-second throughput, Low-latency streaming responses, Support for popular open-source LLMs
- Integrations
- LangChain, LlamaIndex, OpenAI SDK, Vercel AI SDK, Python Requests, JavaScript Fetch API
- Use Cases
- [object Object], [object Object], [object Object], [object Object], [object Object]
- API Available
- Yes
- Tags
- inference, hardware, lpu, ultra-fast-inference, api-provider, llm-hosting, low-latency, real-time-ai, agentic-workflows, semiconductor, openai-compatible
- Added
- 2026-03-17
- Completeness
- 0.95%
Index Score
62.3Adoption
68
Quality
88
Freshness
86
Citations
70
Engagement
0