Quantized Model Information
This repository is a 'FP8-Dynamic' quantized version of
meta-llama/Llama-3.3-70B-Instruct, originally released by Meta AI.
For usage instructions please refer to the original model meta-llama/Llama-3.3-70B-Instruct.
Performance
All benchmarks were done using the LLM Evaluation Harness
| Llama-3.3-70B-Instruct-FP8-Dynamic | Llama-3.3-70B-Instruct (base) | recovery | ||
|---|---|---|---|---|
| mmlu | - | xx | xx | xx |
| xx | xx | xx | ||
| hellaswag | acc | 65.69 | - | |
| acc_sterr | 0.47 | - | ||
| acc_norm | 84.36 | - | ||
| acc_sterr | 0.36 | - |
- Downloads last month
- 1
Model tree for just-add-ai/Llama-3.3-70B-Instruct-FP8-Dynamic
Base model
meta-llama/Llama-3.1-70B
Finetuned
meta-llama/Llama-3.3-70B-Instruct