AI Model Provider Comparison
The provider (e.g., AWS, Azure, Google Cloud) through which you purchase and deploy your LLM directly impacts both performance and cost.
Latency, throughput, and pricing can vary significantly, and not all models are available on every platform. Use this feature to see which providers support the models you need and compare their performance and pricing to optimize the model–provider fit.
Real-time Pricing
Get up-to-date pricing information per million tokens for both input and output across all providers.
Performance Metrics
Compare latency and throughput to find the fastest providers for your specific use case.
Easy Comparison
Sort and filter providers by any metric to quickly identify the best options for your needs.
Providers
Pricing, throughput, and latency for Grok 4:
Provider | Input $/1M | Output $/1M | Latency | Throughput | Updated |
---|---|---|---|---|---|
![]() xAI | $3.00 | $15.00 | ~11s | 32000 tokens/s | 2025-01-12 |
How to Use
Select a Model
Choose from over 80 different AI models including GPT-4, Claude, Gemini, Llama, and more. Each model shows all available providers with their specific pricing and performance metrics.
Compare Providers
View detailed comparisons including input/output pricing per million tokens, latency in milliseconds, and throughput in tokens per second.
Sort & Filter
Click on any column header to sort providers by that metric. Find the most cost-effective option, fastest response times, or highest throughput.
Key Insights
The summary cards at the bottom show quick insights including total providers, lowest pricing, best latency, and highest throughput for the selected model.