{
"service": "InferenceLatency.com",
"providers": [
{
"name": "Google Gemini",
"model": "gemini-2.0-flash-exp",
"latency_ms": null,
"error": "All recent attempts failed",
"success_rate": 0,
"measurements_count": 0
},
{
"name": "Groq",
"model": "llama-3.1-8b-instant",
"latency_ms": 927,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
},
{
"name": "Fireworks AI",
"model": "accounts/fireworks/models/llama-v3p1-8b-instruct",
"latency_ms": null,
"error": "All recent attempts failed",
"success_rate": 0,
"measurements_count": 0
},
{
"name": "Cerebras",
"model": "llama-4-scout-17b-16e-instruct",
"latency_ms": null,
"error": "All recent attempts failed",
"success_rate": 0,
"measurements_count": 0
},
{
"name": "Together AI",
"model": "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
"latency_ms": 548,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
},
{
"name": "Cohere",
"model": "command-a-03-2025",
"latency_ms": 540,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
},
{
"name": "Anthropic",
"model": "claude-3-5-sonnet-20241022",
"latency_ms": null,
"error": "All recent attempts failed",
"success_rate": 0,
"measurements_count": 0
},
{
"name": "OpenRouter",
"model": "mistralai/mistral-7b-instruct",
"latency_ms": 1219,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
},
{
"name": "Hyperbolic",
"model": "meta-llama/Meta-Llama-3.1-8B-Instruct",
"latency_ms": 891,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
},
{
"name": "SambaNova",
"model": "Meta-Llama-3.1-8B-Instruct",
"latency_ms": 280,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
},
{
"name": "OpenAI",
"model": "gpt-4o",
"latency_ms": 1726,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
},
{
"name": "DeepSeek",
"model": "deepseek-chat",
"latency_ms": 2220,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
}
],
"ranked": [
{
"name": "SambaNova",
"model": "Meta-Llama-3.1-8B-Instruct",
"latency_ms": 280,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
},
{
"name": "Cohere",
"model": "command-a-03-2025",
"latency_ms": 540,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
},
{
"name": "Together AI",
"model": "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
"latency_ms": 548,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
},
{
"name": "Hyperbolic",
"model": "meta-llama/Meta-Llama-3.1-8B-Instruct",
"latency_ms": 891,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
},
{
"name": "Groq",
"model": "llama-3.1-8b-instant",
"latency_ms": 927,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
},
{
"name": "OpenRouter",
"model": "mistralai/mistral-7b-instruct",
"latency_ms": 1219,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
},
{
"name": "OpenAI",
"model": "gpt-4o",
"latency_ms": 1726,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
},
{
"name": "DeepSeek",
"model": "deepseek-chat",
"latency_ms": 2220,
"tokens_generated": 1,
"success_rate": 0,
"measurements_count": 0
}
],
"fastest": "SambaNova",
"total_tested": 12,
"successful_tests": 8,
"failed_tests": 4,
"performance_summary": {
"fastest_ms": 280,
"slowest_ms": 2220,
"average_ms": 1044
},
"ai_agent_guidance": {
"recommended_provider": "SambaNova",
"use_case": "Choose 'fastest' provider for lowest latency API calls",
"fallback_order": [
"SambaNova",
"Cohere",
"Together AI",
"Hyperbolic",
"Groq",
"OpenRouter",
"OpenAI",
"DeepSeek"
],
"reliability_score": "8/12 providers responding"
},
"human_readable_summary": "\ud83c\udfc6 Fastest: SambaNova (280ms - Good) | \ud83d\udcca Range: 280-2220ms | \ud83d\udcc8 Average: 1044ms | \u2705 8/12 providers responding",
"timestamp": "2026-01-15T05:05:12.440949Z"
}