# LLM Latency Benchmark Report

**Generated**: 2026-01-27 05:23:21

## Summary

| Model | TTFT (median) | Total (median) | Tokens/sec | Success |
|-------|---------------|----------------|------------|---------|
| openai/gpt-5-nano | N/A | N/A | N/A | 0% |
| openai/gpt-5-mini | N/A | N/A | N/A | 0% |

## Detailed Results

### openai/gpt-5-nano

**All runs failed**

**Individual Runs**

- Run 1: FAILED - litellm.BadRequestError: OpenAIException - Unsupported value: 'reasoning_effort' does not support 'none' with this model. Supported values are: 'minimal', 'low', 'medium', and 'high'.
- Run 2: FAILED - litellm.BadRequestError: OpenAIException - Unsupported value: 'reasoning_effort' does not support 'none' with this model. Supported values are: 'minimal', 'low', 'medium', and 'high'.

### openai/gpt-5-mini

**All runs failed**

**Individual Runs**

- Run 1: FAILED - litellm.BadRequestError: OpenAIException - Unsupported value: 'reasoning_effort' does not support 'none' with this model. Supported values are: 'minimal', 'low', 'medium', and 'high'.
- Run 2: FAILED - litellm.BadRequestError: OpenAIException - Unsupported value: 'reasoning_effort' does not support 'none' with this model. Supported values are: 'minimal', 'low', 'medium', and 'high'.

## Configuration

- **Prompt**: "Explain how a CPU cache works in 3 paragraphs."
- **Max Tokens**: 300
- **Timeout**: 60s
