🚀 AI Latency Test Results

← Back to Home

🏆 Fastest: Groq (426ms - Good) | 📊 Range: 426-1873ms | 📈 Average: 1127ms | ✅ 7/12 providers responding
Full JSON Response (for developers & agents)
{
  "service": "InferenceLatency.com",
  "providers": [
    {
      "name": "Google Gemini",
      "model": "gemini-2.0-flash-exp",
      "latency_ms": null,
      "error": "All recent attempts failed",
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "Groq",
      "model": "llama-3.1-8b-instant",
      "latency_ms": 426,
      "tokens_generated": 1,
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "Fireworks AI",
      "model": "accounts/fireworks/models/llama-v3p1-8b-instruct",
      "latency_ms": null,
      "error": "All recent attempts failed",
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "Cerebras",
      "model": "llama-4-scout-17b-16e-instruct",
      "latency_ms": null,
      "error": "All recent attempts failed",
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "OpenRouter",
      "model": "mistralai/mistral-7b-instruct",
      "latency_ms": null,
      "error": "All recent attempts failed",
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "Together AI",
      "model": "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
      "latency_ms": 463,
      "tokens_generated": 1,
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "SambaNova",
      "model": "Meta-Llama-3.1-8B-Instruct",
      "latency_ms": 480,
      "tokens_generated": 1,
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "Anthropic",
      "model": "claude-3-5-sonnet-20241022",
      "latency_ms": null,
      "error": "All recent attempts failed",
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "Hyperbolic",
      "model": "meta-llama/Meta-Llama-3.1-8B-Instruct",
      "latency_ms": 1262,
      "tokens_generated": 1,
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "OpenAI",
      "model": "gpt-4o",
      "latency_ms": 1672,
      "tokens_generated": 1,
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "DeepSeek",
      "model": "deepseek-chat",
      "latency_ms": 1711,
      "tokens_generated": 1,
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "Cohere",
      "model": "command-a-03-2025",
      "latency_ms": 1873,
      "tokens_generated": 1,
      "success_rate": 0,
      "measurements_count": 0
    }
  ],
  "ranked": [
    {
      "name": "Groq",
      "model": "llama-3.1-8b-instant",
      "latency_ms": 426,
      "tokens_generated": 1,
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "Together AI",
      "model": "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
      "latency_ms": 463,
      "tokens_generated": 1,
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "SambaNova",
      "model": "Meta-Llama-3.1-8B-Instruct",
      "latency_ms": 480,
      "tokens_generated": 1,
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "Hyperbolic",
      "model": "meta-llama/Meta-Llama-3.1-8B-Instruct",
      "latency_ms": 1262,
      "tokens_generated": 1,
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "OpenAI",
      "model": "gpt-4o",
      "latency_ms": 1672,
      "tokens_generated": 1,
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "DeepSeek",
      "model": "deepseek-chat",
      "latency_ms": 1711,
      "tokens_generated": 1,
      "success_rate": 0,
      "measurements_count": 0
    },
    {
      "name": "Cohere",
      "model": "command-a-03-2025",
      "latency_ms": 1873,
      "tokens_generated": 1,
      "success_rate": 0,
      "measurements_count": 0
    }
  ],
  "fastest": "Groq",
  "total_tested": 12,
  "successful_tests": 7,
  "failed_tests": 5,
  "performance_summary": {
    "fastest_ms": 426,
    "slowest_ms": 1873,
    "average_ms": 1127
  },
  "ai_agent_guidance": {
    "recommended_provider": "Groq",
    "use_case": "Choose 'fastest' provider for lowest latency API calls",
    "fallback_order": [
      "Groq",
      "Together AI",
      "SambaNova",
      "Hyperbolic",
      "OpenAI",
      "DeepSeek",
      "Cohere"
    ],
    "reliability_score": "7/12 providers responding"
  },
  "human_readable_summary": "\ud83c\udfc6 Fastest: Groq (426ms - Good) | \ud83d\udcca Range: 426-1873ms | \ud83d\udcc8 Average: 1127ms | \u2705 7/12 providers responding",
  "timestamp": "2026-03-01T22:47:33.108457Z"
}