AI Integrations Directory
85 AI integration connectors and workflow plugins ranked by composite score — covering no-code platforms, API connectors, and agent tool libraries. Each integration is scored on adoption, quality, freshness, citations, and community engagement.
85 integrations
LangChain + OpenAI
by LangChain
Native integration between LangChain and OpenAI's GPT models. Provides seamless access to chat completions, embeddings, and function calling through LangChain's unified interface. Supports streaming, tool use, and structured output via the langchain-openai package.
GitHub Copilot + VS Code
by GitHub
GitHub Copilot integrates into VS Code as a first-party extension, delivering inline ghost-text completions, multi-line suggestions, and a dedicated Copilot Chat panel for conversational refactoring, test generation, and documentation. It leverages Codex and GPT-4 models under the hood, with workspace-aware context from open tabs and the current file.
Meta + HuggingFace (Llama)
by Meta AI
Official Meta Llama model weights distributed through the HuggingFace Hub under Meta's community license. Covers Llama 3.1, 3.2, and 3.3 variants from 1B to 405B parameters with full transformers, TGI, and vLLM compatibility. HuggingFace serves as the primary public distribution channel for Meta's open-weight releases.
LangChain + Anthropic
by LangChain
Official LangChain integration for Anthropic's Claude model family. Exposes Claude's extended context window, vision capabilities, and tool use through LangChain's standard chat model interface. Supports streaming and the full Messages API via the langchain-anthropic package.
Pinecone + OpenAI Embeddings
by Pinecone
Direct integration pairing Pinecone's managed vector database with OpenAI's text-embedding-3 models. Commonly used pattern for production RAG systems where OpenAI generates dense vectors and Pinecone handles ANN retrieval at scale. Supports serverless and pod-based indexes with metadata filtering.
W&B + Hugging Face
by Weights & Biases
Weights & Biases integrates directly into Hugging Face Trainer and PEFT via a built-in report_to callback, logging training loss curves, GPU utilization, gradient norms, and hyperparameters to shareable W&B runs. The integration supports sweep-based hyperparameter optimization and artifact versioning for model checkpoints.
vLLM + NVIDIA
by vLLM Project
vLLM's NVIDIA backend leverages CUDA kernels, FlashAttention-2, and PagedAttention to deliver state-of-the-art throughput for LLM inference on NVIDIA A100, H100, and H200 GPUs. The integration supports tensor and pipeline parallelism across multiple GPUs, FP8/FP16/BF16 quantization, and CUDA graph capture for minimal per-token latency.
LangSmith + LangChain
by LangChain Inc.
LangSmith provides first-class tracing and evaluation for LangChain pipelines, capturing every LLM call, chain step, and tool invocation with full prompt/response payloads. Teams use the integration to debug production failures, build evaluation datasets, and run automated regression tests against golden traces.
OpenAI + Azure OpenAI Service
by Microsoft Azure
Microsoft Azure's managed deployment of OpenAI models including GPT-4o, o1, and DALL-E 3 with enterprise SLAs, private networking, and regional data residency. Provides the same OpenAI API surface with additional Azure IAM, VNet integration, content filtering, and Azure Monitor observability.
LangChain + Pinecone
by LangChain
LangChain VectorStore integration for Pinecone's managed vector database. Enables similarity search, MMR retrieval, and metadata filtering within LangChain RAG pipelines. Supports both serverless and pod-based Pinecone indexes via the langchain-pinecone package.
Cursor + OpenAI
by Anysphere
Cursor is a VS Code fork that uses OpenAI's GPT-4 and o-series models as its reasoning engine for multi-file edits, semantic codebase search, and an agent mode that can autonomously implement features across the entire repository. It offers a Composer panel for multi-file diffs and a codebase-aware chat that indexes the project with embeddings for precise retrieval.
Anthropic + AWS Bedrock
by Amazon Web Services
Anthropic's Claude model family available through Amazon Bedrock's fully managed foundation model service. Provides serverless inference with pay-per-token pricing, AWS IAM authentication, VPC endpoint support, and model evaluation tools. Claude 3.5 Sonnet, Haiku, and Opus are all available through the Bedrock API.
TGI + Hugging Face Hub
by Hugging Face
Text Generation Inference (TGI) by Hugging Face is a production-grade inference server that directly loads models from the Hugging Face Hub via model IDs, handling shard downloading, quantization, and OpenAI-compatible endpoint serving in a single Docker command. It implements continuous batching, speculative decoding, and FlashAttention for optimal throughput on Ampere and Hopper GPUs.
Ollama + Docker
by Ollama
Ollama's official Docker image packages the Ollama runtime for containerized local LLM inference, enabling teams to run quantized GGUF models on CPU or GPU inside Docker Compose stacks or Kubernetes pods. The integration supports GPU passthrough via NVIDIA Container Toolkit and provides an OpenAI-compatible HTTP API for drop-in compatibility with existing tooling.
MCP + GitHub
by Anthropic / GitHub
Official MCP GitHub server providing tools for repository management, issue tracking, pull request review, and code search via the GitHub REST and GraphQL APIs. Enables Claude and other MCP clients to interact with GitHub repositories programmatically without leaving the agent context.
GitHub Copilot + JetBrains
by GitHub
The GitHub Copilot JetBrains plugin brings inline AI completions and Copilot Chat to the entire JetBrains IDE family including IntelliJ IDEA, PyCharm, GoLand, and Rider. It mirrors the VS Code experience with ghost-text suggestions and a side-panel chat, adapting to JetBrains' editor model and keymap conventions.
MCP + Filesystem
by Anthropic
The official Anthropic MCP Filesystem server exposes local file and directory operations to any MCP client. It provides tools for reading, writing, listing, searching, and moving files, enabling Claude and other agents to directly interact with the host filesystem within configurable permission boundaries.
LangChain + Chroma
by LangChain
LangChain VectorStore integration for Chroma, the open-source AI-native embedding database. Ideal for local development and prototyping with zero infrastructure setup. Supports persistent and in-memory collections, metadata filtering, and relevance-scored retrieval via langchain-chroma.
LangChain + Google AI
by LangChain
LangChain integration for Google's AI ecosystem covering both Google AI Studio (Gemini API) and Vertex AI. Supports multimodal inputs, function calling, grounding with Google Search, and long-context processing via the langchain-google-genai and langchain-google-vertexai packages.
Google AI + Vertex AI
by Google Cloud
Google's Gemini and PaLM models served through Vertex AI's managed ML platform with enterprise-grade tooling. Adds model tuning, evaluation pipelines, Model Garden access, Grounding with Google Search, and full GCP IAM/VPC integration on top of the raw Gemini API — the recommended path for production Google AI deployments.
LangChain + HuggingFace
by LangChain
LangChain integration for the HuggingFace ecosystem, covering the Inference API, local transformers pipelines, and HuggingFace Hub embeddings. Enables use of thousands of open-source models within LangChain chains and RAG pipelines via the langchain-huggingface package.
TensorRT-LLM + NVIDIA Triton
by NVIDIA
TensorRT-LLM compiles and optimizes LLMs into fused CUDA kernels using NVIDIA's TensorRT compiler, while the Triton Inference Server backend orchestrates dynamic batching, multi-instance serving, and gRPC/HTTP endpoint management. Together they form NVIDIA's recommended production stack for maximizing tokens-per-second on datacenter GPUs.
LangGraph + LangSmith
by LangChain Inc.
Built-in observability bridge between LangGraph stateful agent graphs and LangSmith's tracing and evaluation platform. Every LangGraph node execution, state transition, and tool call is automatically captured as a structured trace, enabling step-level debugging and regression testing of complex agent workflows.
CrewAI + LangChain
by CrewAI / LangChain
Deep integration allowing CrewAI agents to use the full LangChain tool ecosystem, including web search, code execution, vector store retrieval, and API connectors. CrewAI handles role-based orchestration and task routing while LangChain provides the underlying tool and chain primitives.
Missing an integration?
Submit any AI integration or connector to the index. Our research pipeline scores and enriches it automatically.
Submit an Integration