AI Model Provider Comparison

The provider (e.g., AWS, Azure, Google Cloud) through which you purchase and deploy your LLM directly impacts both performance and cost.

Latency, throughput, and pricing can vary significantly, and not all models are available on every platform. Use this feature to see which providers support the models you need and compare their performance and pricing to optimize the model–provider fit.

Real-time Pricing

Get up-to-date pricing information per million tokens for both input and output across all providers.

Performance Metrics

Compare latency and throughput to find the fastest providers for your specific use case.

Easy Comparison

Sort and filter providers by any metric to quickly identify the best options for your needs.

Providers

Pricing, throughput, and latency for Grok 4:

Provider
Input $/1M
Output $/1M
Latency
Throughput
Updated
xAI logo
xAI
$3.00$15.00~11s32000 tokens/s2025-01-12
Total Providers
1
Lowest Input Price
$3.00
Latency
Varies
Throughput
32000 tokens/s

How to Use

Select a Model

Choose from over 80 different AI models including GPT-4, Claude, Gemini, Llama, and more. Each model shows all available providers with their specific pricing and performance metrics.

Compare Providers

View detailed comparisons including input/output pricing per million tokens, latency in milliseconds, and throughput in tokens per second.

Sort & Filter

Click on any column header to sort providers by that metric. Find the most cost-effective option, fastest response times, or highest throughput.

Key Insights

The summary cards at the bottom show quick insights including total providers, lowest pricing, best latency, and highest throughput for the selected model.