Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Llama2

Benchmarks

Task NameDataset NameSOTA ResultTrend
Attention Operator ThroughputLlama2 7B (32 Q-heads/32 KV-heads/128 Head-dimension)
Attention TFLOPS207.3
30
Jailbreak attackLlama2-7b five finetuned variants
Average ASR0
16
AccuracyLLaMA2-7B zero-shot
Zero-Shot Accuracy67.18
16
Multi-bit WatermarkingLLaMA2-7B 300 tokens (test)
Perplexity7.0486
14
Jailbreak Attackllama2-7b v1 (pretrained)
ASR0
13
LLM TrainingLlama2-70B (64 x H100-8)
Iteration Time (s)7.8
4
LLM TrainingLlama2 7B
Iteration Time (s)1.4
4
QuantizationLLaMA2-7B
Averaged Quantization Time (s)24
4
LLM TrainingLlama2-7B tpu-v5p-512
Iteration Time (s)2.5
3
LLM TrainingLlama2 70B (tpu-v5p-1024)
Iteration Time (s)11.6
2
LLM TrainingLlama2 70B 64 x Trainium2-16
Iteration Time (s)11.2
1
LLM TrainingLlama2 7B 64 x Trainium2-16
Iteration Time (s)1.2
1
LanguageLlama2 7B
Peak Performance2.05
1
Showing 13 of 13 rows