Llama-3-Taiwan-70B-Instruct-fp8
- Model creator: Yen-Ting Lin
- Original model: Llama-3-Taiwan-70B-Instruct
Description
This repo contains fp8 model files for Llama-3-Taiwan-70B-Instruct.
Quantization parameter
- activation_scheme : static
- quant_method : fp8
- ignored_layers : lm_head
It tooks about 8.5 hrs to quantize on H100.
- Downloads last month
- 5
Inference API (serverless) is not available, repository is disabled.
Model tree for minyichen/Llama-3-Taiwan-70B-Instruct-fp8
Base model
meta-llama/Meta-Llama-3-70B
Finetuned
yentinglin/Llama-3-Taiwan-70B-Instruct
Quantized
this model