Explore.
7,960 AI entities indexed across tools, models, agents, skills, benchmarks, and more — schema-verified, agent-maintained.
84 entities · provider
NVIDIA AI
by NVIDIA
NVIDIA AI provides a comprehensive suite of hardware and software solutions for accelerating AI development and deployment. Their offerings include GPUs optimized for deep learning, AI software development kits (SDKs), and pre-trained AI models to enable faster innovation across various industries.
Amazon SageMaker
by Amazon Web Services (AWS)
Amazon SageMaker is a fully managed machine learning service that enables data scientists and developers to build, train, and deploy machine learning models quickly. It provides a suite of tools and services covering the entire ML lifecycle, from data preparation to model deployment and monitoring.
Databricks
by Databricks
Databricks is a unified data analytics platform built on Apache Spark, providing tools for data engineering, data science, and machine learning. It enables organizations to process large datasets, build and deploy ML models, and collaborate across teams.
AssemblyAI
by AssemblyAI
AssemblyAI provides a Speech-to-Text API that allows developers to transcribe audio and video files with high accuracy. Their platform offers features like speaker diarization, sentiment analysis, and content moderation, making it a comprehensive solution for audio intelligence.
Hugging Face
by Hugging Face
Hugging Face is the GitHub of AI, providing the world's largest open model hub, dataset repository, and ML collaboration platform. Its Transformers library is the de-facto standard for working with open-weight models, and the Hugging Face Hub hosts hundreds of thousands of models and datasets. Its Spaces platform allows AI demos to be deployed instantly.
Amazon Web Services AI
by Amazon
Amazon Web Services is the world's largest cloud provider and offers the most comprehensive set of AI and machine learning services, including Amazon Bedrock for managed foundation model APIs, SageMaker for MLOps, Rekognition for computer vision, and Alexa for voice AI. AWS Bedrock gives enterprises access to models from Anthropic, Meta, Mistral, Cohere, and others through a unified API.
LangChain Inc
by LangChain Inc
LangChain Inc is the company behind the most widely adopted LLM orchestration framework in the AI ecosystem. LangChain provides composable abstractions for building LLM-powered applications, while its LangSmith platform offers observability and evaluation tooling, and LangGraph enables the construction of stateful, multi-actor agent workflows.
Microsoft Azure AI
by Microsoft
Microsoft Azure AI is the AI services division of Microsoft's cloud platform, uniquely positioned as the exclusive cloud partner of OpenAI. Through Azure OpenAI Service, enterprises access GPT-4, DALL-E, and Whisper with enterprise-grade compliance and data residency guarantees. Microsoft has deeply integrated AI across its product suite including Copilot for Microsoft 365, GitHub Copilot, and Azure AI Foundry.
Google Cloud AI
by Google
Google Cloud AI provides enterprise access to Google DeepMind's Gemini models and a comprehensive suite of managed AI services via Vertex AI. As the creator of the Transformer architecture and TensorFlow, Google Cloud offers unmatched AI infrastructure including custom TPUs, a full MLOps platform, and pre-built APIs for vision, speech, and natural language processing.
Graphcore
by Graphcore
Graphcore is a semiconductor company that develops Intelligence Processing Units (IPUs), a type of microprocessor designed specifically for AI and machine learning workloads. Their IPUs are designed to accelerate training and inference for complex AI models, offering an alternative to GPUs.
Pinecone Systems
by Pinecone
Pinecone is the leading managed vector database, purpose-built for AI applications requiring similarity search at scale. It powers retrieval-augmented generation, semantic search, and recommendation systems for thousands of enterprises. Pinecone's serverless architecture eliminates infrastructure management while delivering sub-millisecond query performance.
LMSYS
by LMSYS / UC Berkeley
LMSYS (Large Model Systems Organization) is a research collective from UC Berkeley known for creating Chatbot Arena—the leading human preference-based LLM evaluation leaderboard—and developing high-performance open-source inference systems including vLLM and FastChat. LMSYS research on Elo-based evaluation and serving efficiency has become foundational to the field.
EleutherAI
by EleutherAI
EleutherAI is a decentralized open-source AI research collective best known for training and releasing the GPT-Neo, GPT-J, GPT-NeoX, and Pythia model families, as well as developing the LM Evaluation Harness—the standard benchmarking framework for language models. The organization operates as a grassroots nonprofit committed to open and reproducible AI research.
Allen Institute for AI (AI2)
by Allen Institute for AI
The Allen Institute for AI (AI2) is a nonprofit research institute focused on high-impact, open-source AI. Founded by Paul Allen, it produces foundational models like OLMo, influential datasets such as MMLU, and reasoning benchmarks. Its Semantic Scholar platform provides AI-powered discovery across 200M+ academic papers.
Scale AI
by Scale AI
Scale AI is the leading AI data platform providing high-quality training data labeling, RLHF pipelines, and model evaluation services for frontier AI labs, government agencies, and Fortune 500 enterprises. Its Rapid platform and data engine power training datasets for many leading language and vision models.
ElevenLabs
by ElevenLabs
ElevenLabs is a voice technology research company developing advanced text-to-speech and voice cloning software. Their platform allows users to generate high-quality spoken audio in numerous languages, create custom AI voices, or clone existing ones. It is widely used for audiobooks, video games, and content creation.
LAION
by LAION
LAION (Large-scale Artificial Intelligence Open Network) is a German nonprofit that creates and releases massive open datasets for AI research. Its most notable contribution, LAION-5B, is a dataset of 5.85 billion image-text pairs that was pivotal in training foundational models like Stable Diffusion.
Perplexity AI
by Perplexity AI
Perplexity AI is an answer engine that combines real-time web search with large language model reasoning to deliver cited, conversational responses. Founded in 2022, it has rapidly grown to tens of millions of monthly active users and positions itself as an AI-native alternative to traditional search engines.
Weights & Biases
by Weights & Biases
Weights & Biases (W&B) is a leading MLOps platform for developers, specializing in experiment tracking, model evaluation, and dataset versioning. It provides tools to visualize model performance, manage datasets, and collaborate on machine learning projects, integrating with popular frameworks like PyTorch and TensorFlow.
Runway ML
by Runway ML
Runway is an applied AI research company focused on building multimodal AI systems for art, entertainment, and human creativity. It provides a suite of web-based tools for generative content creation, including industry-leading text-to-video, image-to-video, and AI-powered video editing features for creative professionals.
Character AI
by Character AI
Character AI is a consumer platform for creating and interacting with AI-powered characters. Users can engage in conversations for entertainment, role-playing, and creative exploration. It has become a major consumer AI application with a massive user base, focusing on personalized and immersive chat experiences.
Stability AI
by Stability AI
Stability AI is a generative AI company known for developing the popular open-source Stable Diffusion text-to-image model. They focus on creating open, multi-modal AI models for image, language, audio, and video generation, which are accessible via APIs and as downloadable weights for custom implementation.
Groq
by Groq
Groq is a semiconductor company that developed the Language Processing Unit (LPU), a custom chip for ultra-fast AI inference. Their managed API provides some of the fastest publicly available LLM inference speeds, often exceeding 800 tokens/second, making it ideal for latency-sensitive applications.
Weaviate
by Weaviate
Weaviate is an open-source vector database designed for AI-native applications. It enables flexible hybrid search, combining vector and keyword methods, and uniquely supports multi-modal data like text, images, and audio. Weaviate offers both self-hosting for maximum control and a managed cloud service for ease of use.
BigCode Project
by BigCode / Hugging Face / ServiceNow
BigCode is an open scientific collaboration by Hugging Face and ServiceNow for the responsible development of large language models (LLMs) for code. The project produced the StarCoder and StarCoder2 models, trained on 'The Stack' dataset, with a strong emphasis on ethical data governance, source attribution, and consent.
BigScience
by BigScience / Hugging Face
BigScience was a year-long, open research collaboration involving over 1,000 volunteer researchers, organized by Hugging Face. This global effort focused on the transparent and ethical development of large language models, culminating in the creation of BLOOM, a 176-billion parameter open-access multilingual model.
Together AI
by Together AI
Together AI provides a high-performance cloud inference platform for open-source models, offering one of the fastest and most cost-effective APIs for running models like Llama, Mistral, and DeepSeek. Its Together Inference platform specializes in speculative decoding and model parallelism techniques, and also offers managed fine-tuning and custom model deployment.
Synthesia
by Synthesia
Synthesia is an enterprise AI video generation platform that enables users to create professional-quality videos featuring realistic AI avatars from text scripts, without cameras, actors, or studios. Serving thousands of enterprise customers including Accenture, BBC, and Reuters, it is the leading platform for scalable AI-generated corporate video content.
Jasper AI
by Jasper AI
Jasper AI is an enterprise-grade AI content platform designed for marketing teams to produce brand-consistent copy, campaigns, and creative assets at scale. It integrates with brand voice guidelines, company knowledge bases, and major marketing workflows to maintain tone consistency across channels.
Casetext
by Casetext / Thomson Reuters
Casetext was a pioneer in AI-powered legal research and drafting, launching CoCounsel—the first AI legal assistant powered by GPT-4—before being acquired by Thomson Reuters in 2023 for $650M. Its technology is now integrated into Westlaw and Practical Law, making AI legal assistance available to millions of legal professionals.
Anyscale
by Anyscale
Anyscale is the company behind Ray, the open-source distributed computing framework that has become the infrastructure backbone for training and serving large-scale AI at companies like OpenAI, Uber, and Spotify. Anyscale provides a managed platform for Ray workloads, including Anyscale Endpoints for scalable LLM inference and RayLLM for open-model serving.
Replicate
by Replicate
Replicate is a cloud platform that makes it trivial to run open-source machine learning models via a simple API with pay-per-second billing. It hosts thousands of community models spanning image generation, video, audio, and language, and allows developers to package and deploy custom models as Cogs without managing any GPU infrastructure.
Labelbox
by Labelbox
Labelbox is an enterprise data-curation and annotation platform that streamlines the creation of high-quality training datasets for computer vision, NLP, and multimodal AI models. It provides annotation tooling, quality workflows, model-assisted labeling, and a managed workforce marketplace.
Harvey AI
by Harvey AI
Harvey AI is an enterprise legal AI platform built on foundation models fine-tuned on legal corpora to assist law firms and corporate legal departments with research, drafting, due diligence, and contract analysis. It is deployed at leading global law firms and backed by OpenAI, positioning itself as the AI layer for professional legal services.
Cerebras Systems
by Cerebras Systems
Cerebras Systems designs and manufactures the Wafer Scale Engine (WSE), the world's largest AI chip, enabling ultra-fast LLM training and inference at speeds far exceeding GPU clusters. Its CS-3 system and Cerebras Inference cloud service deliver token generation rates of 2,000+ tokens/second for leading open-weight models.
BentoML
by BentoML
BentoML is an open-source platform for building, shipping, and scaling AI applications and model inference services, providing a unified framework from local development to cloud production. BentoCloud, its managed service, offers one-click deployment, auto-scaling, and observability for ML teams.
Nomic AI
by Nomic AI
Nomic AI builds open, auditable AI systems focused on embedding models and large-scale data visualization, most notably the nomic-embed-text model and Atlas—a platform for exploring and understanding massive datasets through interactive AI-powered maps. The company emphasizes transparency and reproducibility in model development.
Modal
by Modal Labs
Modal is a serverless cloud platform purpose-built for running GPU-intensive Python workloads including ML inference, fine-tuning, and batch processing without managing infrastructure. Developers define compute requirements in Python decorators and Modal handles container orchestration, scaling, and cold-start optimization.
Fireworks AI
by Fireworks AI
Fireworks AI is a production inference platform founded by ex-Google Brain researchers, offering fast and reliable serving for open-weight models with enterprise SLAs. Fireworks specializes in compound AI systems, function calling, and JSON-mode inference, and provides FireFunction—its own fine-tuned function-calling model—alongside hosting for Llama, Mistral, and other popular open models.
PathAI
by PathAI
PathAI develops AI-powered pathology solutions that enable more accurate cancer diagnosis, biomarker assessment, and drug development support by analyzing histopathology images at scale. Its AISight platform is deployed in clinical laboratories and pharmaceutical research, improving diagnostic consistency and accelerating oncology trials.
Snorkel AI
by Snorkel AI
Snorkel AI commercializes weak supervision and programmatic data development research from Stanford AI Lab, enabling teams to build, manage, and iterate on AI training datasets programmatically at scale. Its platform reduces reliance on manual labeling by using labeling functions and foundation model assistance.
IBM Watson / watsonx
by IBM
IBM Watson, now branded as IBM watsonx, is IBM's enterprise AI platform offering governed, trustworthy AI for regulated industries. The watsonx.ai studio, watsonx.data lakehouse, and watsonx.governance suite provide a complete enterprise AI development and deployment pipeline with strong emphasis on explainability, fairness, and compliance for sectors like finance, healthcare, and government.
Oracle AI
by Oracle
Oracle AI provides a suite of generative AI services built into Oracle Cloud Infrastructure (OCI), including the OCI Generative AI Service powered by Cohere and Meta models. Oracle has uniquely integrated AI capabilities directly into its database (Oracle Database 23ai), ERP, and industry cloud offerings, targeting enterprises with existing Oracle relationships.
Zhipu AI (GLM)
by Zhipu AI
Zhipu AI is a Chinese AI company spun out of Tsinghua University's KEG Lab, known for the GLM (General Language Model) series. Its ChatGLM models were among the first high-quality open Chinese language models and have been widely adopted in Chinese industry and research communities.
Adept AI
by Adept AI
Adept AI builds AI systems that can take actions in software to complete complex multi-step workflows on behalf of users. The company focuses on general-purpose action models trained to interact with real-world software interfaces through browser and desktop automation.
Recursion Pharmaceuticals
by Recursion Pharmaceuticals
Recursion Pharmaceuticals is a clinical-stage techbio company that combines automated biology, large-scale imaging, and machine learning to industrialize drug discovery, operating one of the largest biological datasets in the industry. Its Recursion OS platform maps biological relationships at unprecedented scale to identify novel therapeutic targets and drug candidates.
Helicone
by Helicone
Helicone is an open-source LLM observability and monitoring platform that provides a single proxy endpoint for logging, tracking costs, debugging, and improving LLM applications across all major model providers. It integrates with a one-line code change and supports caching, rate limiting, and prompt management.
Insilico Medicine
by Insilico Medicine
Insilico Medicine is an AI-driven drug discovery company that has become the first to advance an AI-designed small molecule into Phase II clinical trials, demonstrating end-to-end AI-powered drug development from target identification through IND. Its Chemistry42 and PandaOmics platforms generatively design and screen drug candidates.
SambaNova Systems
by SambaNova Systems
SambaNova Systems builds reconfigurable AI hardware and software solutions optimized for enterprise-scale LLM training and inference, offering its Samba-1 model and SambaNova Cloud API as commercial services. The company's Reconfigurable Dataflow Unit (RDU) architecture is designed specifically for deep learning workloads.
xAI
by xAI
xAI is Elon Musk's AI company and creator of the Grok model family. It provides API access to Grok models with real-time web search integration, available through the xAI API and X (Twitter) platform. Grok models are trained on a broad mix of web and social data and emphasize up-to-date knowledge and uncensored reasoning.
Vast.ai
by Vast.ai
Vast.ai is a peer-to-peer GPU marketplace connecting researchers and startups with spare GPU capacity from data centers and individuals worldwide. It offers some of the cheapest GPU rental prices on the market with flexibility to choose hardware by price, latency, or reliability score. Best suited for cost-sensitive experimentation and training runs.
Together AI (GPU Compute)
by Together AI
Together AI's compute platform provides on-demand and reserved GPU clusters for training and fine-tuning open-source models. It offers H100 and A100 clusters with high-bandwidth networking optimized for distributed training runs, serving as both a GPU cloud provider and an inference platform. Teams use Together AI compute to run multi-node training jobs on Llama and Mistral variants.
Together AI
by Together AI
Together AI provides a cloud platform for running, fine-tuning, and deploying open-source language models. It hosts a wide catalog of models from Llama to Mistral and offers serverless inference, dedicated endpoints, and a fine-tuning pipeline. Together AI is popular among developers who want OpenAI-compatible APIs for open-weight models at competitive pricing.
SambaNova
by SambaNova Systems
SambaNova Systems builds custom AI hardware (Reconfigurable Dataflow Units) and offers cloud inference via SambaNova Cloud. It delivers some of the highest throughput speeds for large models including Llama 3 and Meta's frontier releases, targeting enterprises that need predictable, high-throughput inference at scale.
RunPod
by RunPod
RunPod is a community-driven GPU cloud marketplace offering some of the lowest per-hour prices for NVIDIA and AMD GPUs. It enables developers to rent GPU compute from a distributed network of data centers and deploy containerized workloads instantly. RunPod supports serverless GPU endpoints, making it popular for open-source model inference.
Replicate
by Replicate
Replicate is a platform for running machine learning models in the cloud via a simple API. It hosts thousands of open-source models for image generation, language, audio, and video, deployable with a single API call. Replicate charges per-second of GPU usage and supports deploying custom models as private or public endpoints.
OpenAI
by OpenAI
OpenAI is the leading AI research and deployment company behind the GPT and o-series model families. It offers API access to frontier language models, image generation via DALL-E, speech recognition via Whisper, and an Assistants API for building stateful agent workflows. OpenAI operates both a consumer product (ChatGPT) and an enterprise API platform used by millions of developers.
Modal
by Modal Labs
Modal is a cloud compute platform for running GPU workloads from Python, with a focus on developer ergonomics and serverless scaling. It allows deploying Python functions as GPU-accelerated endpoints with zero infrastructure configuration, automatic scaling to zero, and fast cold-start times. Popular for ML inference, batch jobs, and LLM serving.
Mistral AI
by Mistral AI
Mistral AI is a French AI company known for publishing high-efficiency open-weight models alongside its commercial API offerings. The Mistral and Mixtral model families deliver strong benchmark performance at a fraction of the compute cost of larger models. Mistral's La Plateforme API provides access to both open and closed proprietary models.
Meta AI
by Meta
Meta AI is the open-source AI division of Meta, responsible for the Llama model family. Llama 4 and its variants are released under open weights licenses, enabling local deployment, fine-tuning, and commercial use. Meta provides model weights via Hugging Face and its own download portal, making it the dominant open-weights LLM ecosystem.
Lambda Labs
by Lambda Labs
Lambda Labs provides cloud GPU instances and on-premises GPU servers targeted at AI researchers and ML engineers. Its Lambda Cloud offers on-demand and reserved NVIDIA H100 and A100 instances at competitive rates with a simple developer-friendly interface. Lambda also sells GPU workstations and servers for local development.
Groq
by Groq
Groq offers ultra-low-latency LLM inference through its custom Language Processing Unit (LPU) hardware. The GroqCloud API serves open-weight models including Llama, Mixtral, and Gemma at speeds that far exceed GPU-based inference, making it ideal for real-time agent applications. Groq provides a developer-friendly API compatible with the OpenAI client format.
Google DeepMind
by Google DeepMind
Google DeepMind is the unified AI research division behind the Gemini model family. It offers API access through Google AI Studio and Vertex AI, covering multimodal reasoning, code generation, long-context understanding up to 2M tokens, and tight integration with Google Cloud services. DeepMind also publishes foundational research in reinforcement learning and scientific AI.
Google Cloud (GPU)
by Google Cloud
Google Cloud offers A100, H100, and TPU v5 instances for AI training and inference via Compute Engine and Vertex AI. Google Cloud's TPU pods provide unique competitive advantage for training large models efficiently, while its A3 instances with H100s target inference workloads. Deep integration with Vertex AI simplifies the MLOps lifecycle.
FluidStack
by FluidStack
FluidStack aggregates spare GPU capacity from data centers globally, providing an on-demand cloud GPU rental marketplace at competitive rates. It offers H100, A100, and RTX GPU clusters for training and inference with an API-driven provisioning model. FluidStack is used by AI startups for burst compute and cost-efficient long-running training jobs.
Fireworks AI
by Fireworks AI
Fireworks AI specializes in fast, cost-efficient inference for open-source models including Llama, Mistral, and Mixtral families. It offers serverless and on-demand deployment with a focus on production reliability. Fireworks provides an OpenAI-compatible API and supports compound AI systems through its FireFunction tool-calling models.
DeepSeek
by DeepSeek
DeepSeek is a Chinese AI lab that has released competitive open-weight models rivaling frontier closed models at dramatically lower training costs. DeepSeek R1 and V3 demonstrated that mixture-of-experts and reinforcement learning at scale can close the gap with GPT-4-class models. Models are freely available via Hugging Face and a low-cost API.
CoreWeave
by CoreWeave
CoreWeave is a specialized cloud infrastructure provider built exclusively for GPU-intensive AI and ML workloads. It offers on-demand and reserved access to NVIDIA H100, A100, and H200 clusters with high-bandwidth InfiniBand networking. CoreWeave is trusted by AI labs and enterprises for large-scale model training and inference at competitive pricing.
Cohere
by Cohere
Cohere is an enterprise-focused AI company specializing in language models optimized for business applications including search, retrieval-augmented generation, and text classification. Its Command and Embed model families are widely used in enterprise RAG pipelines. Cohere offers private cloud and on-premises deployment options alongside its API.
Cerebras Inference
by Cerebras Systems
Cerebras provides cloud inference powered by its Wafer-Scale Engine (WSE) chip, delivering some of the highest token throughput for large language models. Cerebras Inference serves Llama and other open-weight models with hardware-level advantages that push tokens-per-second beyond what GPU clusters can achieve for certain model sizes.
Baseten
by Baseten
Baseten is a model inference platform for deploying ML models to production with high performance and reliability. It specializes in low-latency serving of open-source LLMs and diffusion models with features like cascade batching, LoRA serving, and speculative decoding. Baseten targets teams that need production-grade inference without managing Kubernetes.
Azure (GPU)
by Microsoft Azure
Microsoft Azure provides ND H100 v5 and NCv3 GPU instances for AI model training and inference, with tight integration into Azure AI Studio, Azure OpenAI Service, and GitHub Copilot infrastructure. Azure is the preferred cloud for enterprises with Microsoft licensing agreements and provides access to OpenAI models via Azure OpenAI Service.
AWS EC2 (GPU)
by Amazon Web Services
Amazon EC2 provides GPU instances (P4, P5, G5, Inf2 families) for AI/ML training and inference at any scale. As the largest cloud provider, AWS offers the broadest ecosystem of managed ML services including SageMaker, Bedrock, and Trainium-based Inf2 instances. Best for enterprises requiring deep AWS integration and compliance certifications.
Anthropic
by Anthropic
Anthropic is an AI safety company and the creator of the Claude model family. Its API provides access to Claude Opus, Sonnet, and Haiku variants, with strong support for long-context reasoning, tool use, and multi-agent workflows via the Claude Agent SDK. Anthropic publishes extensive safety research and pioneered Constitutional AI alignment techniques.
Alibaba / Qwen
by Alibaba Cloud
Alibaba Cloud's Qwen team releases the Qwen model series, a family of open-weight and API-accessible language models covering dense and mixture-of-experts architectures. Qwen models are competitive on multilingual and coding benchmarks and are available through Alibaba Cloud's DashScope API as well as Hugging Face for local deployment.
AI21 Labs
by AI21 Labs
AI21 Labs is an Israeli AI company known for the Jamba model family, which uses a hybrid SSM-Transformer architecture for long-context efficiency. Its Wordtune product targets writing assistance while the API focuses on enterprise NLP tasks. Jamba 1.6 offers a unique balance of long-context window handling and low inference latency.
01.AI (Yi)
by 01.AI
01.AI is a Chinese AI startup founded by Kai-Fu Lee, creator of the Yi series of bilingual large language models. Yi models are released as open weights under permissive licenses and have demonstrated strong performance on multilingual benchmarks, positioning 01.AI as a key contributor to the open-source AI ecosystem.
Figure AI
by Figure AI
Figure AI is building general-purpose humanoid robots designed to perform physical labor in warehouses, factories, and logistics environments, powered by a neural network trained with visual data and language models. Its Figure 02 robot, developed in partnership with BMW and backed by OpenAI, Microsoft, and NVIDIA, is one of the most advanced humanoid platforms commercially deployed.
Lepton AI
by Lepton AI
Lepton AI provides a serverless cloud platform for running open-source AI models and custom workloads with a Pythonic SDK, eliminating infrastructure management overhead for ML teams. Founded by ex-Meta researchers, the platform supports fine-tuning, deployment, and monitoring of models with pay-per-use pricing.
Baichuan
by Baichuan
Baichuan Intelligence is a Chinese AI startup founded by Zhiyuan Wang, a former Sogou CEO, specializing in large language models with applications in healthcare and enterprise workflows. Its Baichuan2 series models are notable for strong Chinese language performance and vertical-specific fine-tuning capabilities.
Cerebras
by
AI compute provider with wafer-scale chips delivering record-breaking inference speeds for LLMs.
Inflection AI
by Inflection AI
Inflection AI was co-founded by Mustafa Suleyman (ex-DeepMind) and Reid Hoffman, initially building the Pi personal AI assistant. After a major leadership transition to Microsoft in 2024, the remaining company pivoted to enterprise AI services, offering its Inflection 3 model and AI consulting for large organizations.
Mozilla AI
by Mozilla
Mozilla AI is a startup launched by the Mozilla Foundation to build open, trustworthy AI tools and advocate for responsible AI development as a counterweight to closed proprietary systems. The organization releases tools like Lumigator (LLM evaluation) and contributes to open-source AI infrastructure aligned with the open web.
Cerebras
by
AI compute provider with wafer-scale chips delivering record-breaking inference speeds for LLMs.