AI Hosting & Inference

Platforms for deploying and serving AI models

22 tools

Anyscale

Paid

Production platform for Ray-based AI applications. Scale from development to production seamlessly.

AI Hosting & Inference

AWS Bedrock

Paid

Managed service for foundation models. Access Claude, Llama, Mistral, and more via unified AWS API.

AI Hosting & Inference

Azure OpenAI Service

Paid

OpenAI models on Azure infrastructure. Enterprise security, compliance, and regional deployment options.

AI Hosting & Inference

Baseten

Paid

Inference platform for ML models. Deploy custom models with auto-scaling and GPU optimization.

AI Hosting & Inference

Cerebras Inference

Paid

Wafer-scale chip inference. Extremely fast token generation with custom hardware architecture.

AI Hosting & Inference

Cloudflare Workers AI

Freemium

Run AI models on Cloudflare's edge network. Low-latency inference close to users. Free tier included.

AI Hosting & Inference

CoreWeave

Paid

GPU cloud provider specializing in AI workloads. Large-scale GPU clusters for training and inference.

AI Hosting & Inference

Fireworks AI

Paid

Fast inference with function calling and JSON mode. Optimized for production AI applications.

AI Hosting & Inference

Fly.io

Freemium

Deploy apps close to users globally. GPU machines available. Good for low-latency AI inference at edge.

AI Hosting & Inference

Google Vertex AI

Paid

Google Cloud's ML platform. Access Gemini models, AutoML, and custom model training/serving.

AI Hosting & Inference

Groq

Freemium

Ultra-fast LLM inference on custom LPU hardware. Fastest token generation for Llama and Mistral models.

AI Hosting & Inference

Hugging Face Inference

Freemium

Serverless API for 300K+ models on Hugging Face Hub. Free tier for popular models, dedicated endpoints for production.

AI Hosting & Inference

Lambda Labs

Paid

GPU cloud for deep learning. On-demand A100/H100 instances. Simple pricing, developer-friendly.

AI Hosting & Inference

Lepton AI

Paid

AI inference platform with OpenAI-compatible API. Fast model deployment with built-in monitoring.

AI Hosting & Inference

Modal

Paid

Serverless cloud for AI/ML. Run GPU workloads without managing infrastructure. Pay per second.

AI Hosting & Inference

OpenRouter

Paid

Unified API gateway to 100+ LLMs. Single API key for OpenAI, Anthropic, Google, Meta, and more.

AI Hosting & Inference

Railway

Freemium

Modern PaaS for deploying any stack. One-click deploys, auto-scaling, built-in Postgres. Great for AI app backends.

AI Hosting & Inference

Replicate

Paid

Run open-source models via API. Pay per second of compute. Large model zoo with one-click deployment.

AI Hosting & Inference

RunPod

Paid

GPU cloud for AI inference and training. On-demand and spot GPUs. Serverless endpoint option.

AI Hosting & Inference

Supabase

Freemium

Open-source Firebase alternative. PostgreSQL + Auth + Storage + Realtime + Edge Functions in one platform.

AI Hosting & Inference

Together AI

Paid

Fast inference for open models. Competitive pricing on Llama, Mistral, and other open-source models.

AI Hosting & Inference

Vercel

Freemium

Frontend cloud platform. Best for Next.js. Automatic preview deployments, edge network, serverless functions.

AI Hosting & Inference