apt install git-lfs
git lfs install
git clone https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct
sudo apt-get update && sudo apt-get -y install python3.10 python3-pip openmpi-bin libopenmpi-dev
pip3 install tensorrt_llm==0.13.0 --extra-index-url https://pypi.nvidia.com
git clone -b v0.13.0 https://github.com/NVIDIA/TensorRT-LLM.git
int4 awq:
python TensorRT-LLM/examples/quantization/quantize.py --model_dir ./Meta-Llama-3.1-8B-Instruct --dtype float16 --qformat int4_awq --batch_size 64 --awq_block_size 128 --output_dir ./tllm_checkpoint_1gpu_int4_awq --calib_size 32
trtllm-build --checkpoint_dir ./tllm_checkpoint_1gpu_int4_awq --output_dir ./tmp/llama/8B/trt_engines/int4_awq/1-gpu --gpt_attention_plugin auto --gemm_plugin auto --max_num_tokens 65536 --max_input_len 1048576 --max_batch_size 64 --gather_generation_logits
Upload
huggingface-cli upload ss-galileo/llama3.1-8b ./tmp/llama/8B/trt_engines/int4_awq/1-gpu/rank0.engine rank0.engine
huggingface-cli upload ss-galileo/llama3.1-8b ./tmp/llama/8B/trt_engines/int4_awq/1-gpu/config.json config.json
and the "tokenizer.json", "tokenizer_config.json", "special_tokens_map.json" from meta-llama/Meta-Llama-3.1-8B-Instruct
- Downloads last month
- 1
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
HF Inference API was unable to determine this model's library.