Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Llama3

Benchmarks

Task NameDataset NameSOTA ResultTrend
Multi-bit LLM WatermarkingLLaMA3-8B-Base Max 256 Tokens
AUC1
20
Multi-bit LLM WatermarkingLLaMA3-8B-Base Max 128 Tokens
AUC1
20
Jailbreak attackLlama3-8b
Average ASR0
16
Jailbreak Attackllama3-8b pretrained v1
ASR0
13
Defending against gradient-based attacksLlama3 AutoDAN Attack (test)
ASR10.57
10
Training ThroughputLlama3 8B (train)
Throughput (128K SeqLen)2,320.47
5
Training Memory Usage ProfilingLlama3-8B 8×H100s
Peak Memory Usage (128K)21.1
5
QuantizationLLaMA3-8B
Averaged Quantization Time (s)27
4
In-Context LearningLlama3-8B Scenario 5 ICL prompts
Accuracy82.8
3
In-Context LearningLlama3-8B Scenario 2: More Pos. (P={-5,...,-1})
Accuracy78.18
3
In-Context LearningLlama3-8B Baseline (P={-1}, L={14})
Accuracy78.65
3
Training EfficiencyLlama3-8x70B Coarse-grained
MFU41.6
3
Model CompressionLlama3-1b
Energy Consumed (kWh)0.0765
2
Showing 13 of 13 rows