Inference Metrics

Loading…

Failed to load metrics. Check the API connection.

Total Requests

across all models

Throughput

requests / second avg

Tokens Processed

total tokens served

Error Rate

of all requests

Request Volume

Hourly request count by model

Live

Throughput

Requests per second

Latency Distribution

Average p50 / p95 / p99 over window

p50 Median
p95 Tail
p99 Extreme

Model Breakdown

Per-model performance summary

Model Requests p50 ms p95 ms Errors
Loading…