{"service":"InferenceLatency.com","endpoint":"custom-benchmark","description":"Custom prompt benchmarking with user-specified parameters","test_parameters":{"prompt":"Hi","max_tokens":1,"providers_requested":"all","prompt_length":2},"results":[{"name":"OpenAI","latency_ms":687,"tokens_generated":1,"tokens_per_second":1.46,"prompt_length":2,"custom_prompt":"Hi"}],"fastest_provider":"OpenAI","total_tested":1,"methodology":"Custom prompt benchmarking with specified parameters","provenance":{"last_checked":"2026-05-17T05:09:18.625427Z","methodology_version":"v1.2","api_version":"v1.0"},"timestamp":"2026-05-17T05:09:18.625435Z"}