📊 Reliability Metrics Analysis

← Back to Home

🔧 Reliability Analysis: 8 providers | 📊 Availability: 100.0% | ✅ SLA compliance: 75.0%
Full JSON Response (for developers & agents)
{
  "service": "InferenceLatency.com",
  "endpoint": "/reliability-metrics",
  "timestamp": "2025-09-07T23:19:17.099557Z",
  "provider_reliability": {
    "OpenAI": {
      "current_performance": {
        "latency_ms": 309,
        "model": "GPT-4o",
        "status": "operational"
      },
      "percentile_analysis": {
        "p50": 433,
        "p95": 1097,
        "p99": 4957,
        "sample_size": 100
      },
      "reliability_metrics": {
        "availability_percent": 100.0,
        "error_rate_percent": 0.0,
        "sla_compliance_percent": 100.0,
        "requests_24h": 9,
        "mean_time_to_respond": 594.11
      },
      "quality_grade": "A+"
    },
    "Groq": {
      "current_performance": {
        "latency_ms": 1956,
        "model": "llama-3.1-8b-instant",
        "status": "operational"
      },
      "percentile_analysis": {
        "p50": 2061,
        "p95": 19208,
        "p99": 34027,
        "sample_size": 100
      },
      "reliability_metrics": {
        "availability_percent": 100.0,
        "error_rate_percent": 0.0,
        "sla_compliance_percent": 33.33,
        "requests_24h": 9,
        "mean_time_to_respond": 4955.33
      },
      "quality_grade": "D"
    },
    "Claude": {
      "current_performance": {
        "latency_ms": 1065,
        "model": "Claude Sonnet 4",
        "status": "operational"
      },
      "percentile_analysis": {
        "p50": 1143,
        "p95": 4595,
        "p99": 40662,
        "sample_size": 100
      },
      "reliability_metrics": {
        "availability_percent": 100.0,
        "error_rate_percent": 0.0,
        "sla_compliance_percent": 44.44,
        "requests_24h": 9,
        "mean_time_to_respond": 7426.11
      },
      "quality_grade": "D"
    },
    "OpenRouter": {
      "current_performance": {
        "latency_ms": 998,
        "model": "Mistral",
        "status": "operational"
      },
      "percentile_analysis": {
        "p50": 1184,
        "p95": 18714,
        "p99": 33061,
        "sample_size": 100
      },
      "reliability_metrics": {
        "availability_percent": 100.0,
        "error_rate_percent": 0.0,
        "sla_compliance_percent": 77.78,
        "requests_24h": 9,
        "mean_time_to_respond": 2912.0
      },
      "quality_grade": "D"
    },
    "Google Gemini": {
      "current_performance": {
        "latency_ms": 304,
        "model": "Gemini-2.0-Flash",
        "status": "operational"
      },
      "percentile_analysis": {
        "p50": 366,
        "p95": 840,
        "p99": 1533,
        "sample_size": 86
      },
      "reliability_metrics": {
        "availability_percent": 100.0,
        "error_rate_percent": 0.0,
        "sla_compliance_percent": 100.0,
        "requests_24h": 9,
        "mean_time_to_respond": 309.78
      },
      "quality_grade": "A+"
    },
    "Together AI": {
      "current_performance": {
        "latency_ms": 471,
        "model": "Llama3.1-8B-Turbo",
        "status": "operational"
      },
      "percentile_analysis": {
        "p50": 389,
        "p95": 18094,
        "p99": 32963,
        "sample_size": 100
      },
      "reliability_metrics": {
        "availability_percent": 100.0,
        "error_rate_percent": 0.0,
        "sla_compliance_percent": 77.78,
        "requests_24h": 9,
        "mean_time_to_respond": 2354.11
      },
      "quality_grade": "D"
    },
    "Fireworks AI": {
      "current_performance": {
        "latency_ms": 2377,
        "model": "Llama3.1-8B",
        "status": "operational"
      },
      "percentile_analysis": {
        "p50": 505,
        "p95": 18563,
        "p99": 32993,
        "sample_size": 100
      },
      "reliability_metrics": {
        "availability_percent": 100.0,
        "error_rate_percent": 0.0,
        "sla_compliance_percent": 66.67,
        "requests_24h": 9,
        "mean_time_to_respond": 2570.33
      },
      "quality_grade": "D"
    },
    "HF GPT OSS 120B (Cerebras)": {
      "current_performance": {
        "latency_ms": 335,
        "model": "GPT OSS 120B",
        "status": "operational"
      },
      "percentile_analysis": null,
      "reliability_metrics": {
        "availability_percent": 100,
        "error_rate_percent": 0,
        "sla_compliance_percent": 100,
        "requests_24h": 0,
        "mean_time_to_respond": null
      },
      "quality_grade": "A+"
    }
  },
  "platform_summary": {
    "overall_availability_percent": 100.0,
    "overall_sla_compliance_percent": 75.0,
    "total_providers_analyzed": 8,
    "grade_distribution": {
      "A+": 3,
      "A": 0,
      "B": 0,
      "C": 0,
      "D": 5
    }
  },
  "sla_targets": {
    "availability_target": "99.5%",
    "latency_target": "95% of requests under 2000ms",
    "error_rate_target": "< 0.5%"
  },
  "methodology": {
    "data_window": "24 hours",
    "percentile_calculations": "Based on last 100 successful requests",
    "grading_criteria": "Industry-standard SLA benchmarks"
  }
}