Groq + LangChain
by Groq · freemium · Last verified 2026-03-17
LangChain chat model integration for Groq's Language Processing Unit (LPU) inference API. Enables ultra-low-latency LLM calls within LangChain chains and agents with first-token latency under 100ms. Supports Llama 3, Mixtral, and Gemma models served on Groq hardware via the langchain-groq package.
https://python.langchain.com/docs/integrations/chat/groq ↗C+
C+—Average
Adoption: BQuality: AFreshness: A+Citations: C+Engagement: F
Specifications
- License
- MIT
- Pricing
- freemium
- Capabilities
- ultra-low-latency, openai-compatible-api, streaming, function-calling, lpu-acceleration
- Integrations
- groq, langchain
- Use Cases
- real-time-ai-apps, low-latency-agents, interactive-chatbots, voice-ai
- API Available
- Yes
- Tags
- groq, langchain, fast-inference, lpu, low-latency
- Added
- 2026-03-17
- Completeness
- 100%
Index Score
57.4Adoption
65
Quality
87
Freshness
90
Citations
56
Engagement
0
Put AI to work for your business
Deploy this integration alongside autonomous AaaS agents that handle tasks end-to-end — no babysitting required.