Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Long-Context LLM Inference

Benchmarks

Task NameDataset NameSOTA ResultTrend
LLM InferenceLong-Context LLM Inference Decode
Latency (ms)0.13
8
LLM InferenceLong-Context LLM Inference (Prefill)
Prefill Latency (ms)0.62
6
Showing 2 of 2 rows