AI Hosting & Inference
Platforms for deploying and serving AI models
22 tools
Anyscale
PaidProduction platform for Ray-based AI applications. Scale from development to production seamlessly.
AI Hosting & InferenceAWS Bedrock
PaidManaged service for foundation models. Access Claude, Llama, Mistral, and more via unified AWS API.
AI Hosting & InferenceAzure OpenAI Service
PaidOpenAI models on Azure infrastructure. Enterprise security, compliance, and regional deployment options.
AI Hosting & InferenceBaseten
PaidInference platform for ML models. Deploy custom models with auto-scaling and GPU optimization.
AI Hosting & InferenceCerebras Inference
PaidWafer-scale chip inference. Extremely fast token generation with custom hardware architecture.
AI Hosting & InferenceCloudflare Workers AI
FreemiumRun AI models on Cloudflare's edge network. Low-latency inference close to users. Free tier included.
AI Hosting & InferenceCoreWeave
PaidGPU cloud provider specializing in AI workloads. Large-scale GPU clusters for training and inference.
AI Hosting & InferenceFireworks AI
PaidFast inference with function calling and JSON mode. Optimized for production AI applications.
AI Hosting & InferenceFly.io
FreemiumDeploy apps close to users globally. GPU machines available. Good for low-latency AI inference at edge.
AI Hosting & InferenceGoogle Vertex AI
PaidGoogle Cloud's ML platform. Access Gemini models, AutoML, and custom model training/serving.
AI Hosting & InferenceGroq
FreemiumUltra-fast LLM inference on custom LPU hardware. Fastest token generation for Llama and Mistral models.
AI Hosting & InferenceHugging Face Inference
FreemiumServerless API for 300K+ models on Hugging Face Hub. Free tier for popular models, dedicated endpoints for production.
AI Hosting & InferenceLambda Labs
PaidGPU cloud for deep learning. On-demand A100/H100 instances. Simple pricing, developer-friendly.
AI Hosting & InferenceLepton AI
PaidAI inference platform with OpenAI-compatible API. Fast model deployment with built-in monitoring.
AI Hosting & InferenceModal
PaidServerless cloud for AI/ML. Run GPU workloads without managing infrastructure. Pay per second.
AI Hosting & InferenceOpenRouter
PaidUnified API gateway to 100+ LLMs. Single API key for OpenAI, Anthropic, Google, Meta, and more.
AI Hosting & InferenceRailway
FreemiumModern PaaS for deploying any stack. One-click deploys, auto-scaling, built-in Postgres. Great for AI app backends.
AI Hosting & InferenceReplicate
PaidRun open-source models via API. Pay per second of compute. Large model zoo with one-click deployment.
AI Hosting & InferenceRunPod
PaidGPU cloud for AI inference and training. On-demand and spot GPUs. Serverless endpoint option.
AI Hosting & InferenceSupabase
FreemiumOpen-source Firebase alternative. PostgreSQL + Auth + Storage + Realtime + Edge Functions in one platform.
AI Hosting & InferenceTogether AI
PaidFast inference for open models. Competitive pricing on Llama, Mistral, and other open-source models.
AI Hosting & InferenceVercel
FreemiumFrontend cloud platform. Best for Next.js. Automatic preview deployments, edge network, serverless functions.
AI Hosting & Inference