{"service":"InferenceLatency.com","endpoint":"custom-benchmark","description":"Custom prompt benchmarking with user-specified parameters","test_parameters":{"prompt":"hello","max_tokens":5,"providers_requested":"all","prompt_length":5},"results":[{"name":"OpenAI","latency_ms":513,"tokens_generated":4,"tokens_per_second":7.8,"prompt_length":5,"custom_prompt":"hello"}],"fastest_provider":"OpenAI","total_tested":1,"methodology":"Custom prompt benchmarking with specified parameters","provenance":{"last_checked":"2026-05-05T12:12:57.008378Z","methodology_version":"v1.2","api_version":"v1.0"},"timestamp":"2026-05-05T12:12:57.008386Z"}