| Task Name | Dataset Name | SOTA Result | Trend | |
|---|---|---|---|---|
| Attention Operator Throughput | Llama2 7B (32 Q-heads/32 KV-heads/128 Head-dimension) | Attention TFLOPS207.3 | 30 | |
| Jailbreak attack | Llama2-7b five finetuned variants | Average ASR0 | 16 | |
| Accuracy | LLaMA2-7B zero-shot | Zero-Shot Accuracy67.18 | 16 | |
| Multi-bit Watermarking | LLaMA2-7B 300 tokens (test) | Perplexity7.0486 | 14 | |
| Jailbreak Attack | llama2-7b v1 (pretrained) | ASR0 | 13 | |
| LLM Training | Llama2-70B (64 x H100-8) | Iteration Time (s)7.8 | 4 | |
| LLM Training | Llama2 7B | Iteration Time (s)1.4 | 4 | |
| Quantization | LLaMA2-7B | Averaged Quantization Time (s)24 | 4 | |
| LLM Training | Llama2-7B tpu-v5p-512 | Iteration Time (s)2.5 | 3 | |
| LLM Training | Llama2 70B (tpu-v5p-1024) | Iteration Time (s)11.6 | 2 | |
| LLM Training | Llama2 70B 64 x Trainium2-16 | Iteration Time (s)11.2 | 1 | |
| LLM Training | Llama2 7B 64 x Trainium2-16 | Iteration Time (s)1.2 | 1 | |
| Language | Llama2 7B | Peak Performance2.05 | 1 |