Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
astronomer
/
Llama-3-8B-GPTQ-4-Bit
like
6
Follow
Astronomer
5
Text Generation
Transformers
Safetensors
wikitext
llama
llama-3
facebook
meta
astronomer
gptq
pretrained
quantized
finetuned
Inference Endpoints
text-generation-inference
4-bit precision
arxiv:
2210.17323
License:
llama-3
Model card
Files
Files and versions
Community
1
Train
Deploy
Use this model
72d7127
Llama-3-8B-GPTQ-4-Bit
1 contributor
History:
6 commits
davidxmle
Update README.md
72d7127
verified
9 months ago
.gitattributes
1.52 kB
initial commit
9 months ago
LICENSE.txt
7.8 kB
Upload Llama 3 license, use policy and modified generation_config
9 months ago
README.md
6.11 kB
Update README.md
9 months ago
USE_POLICY.md
4.7 kB
Upload Llama 3 license, use policy and modified generation_config
9 months ago
config.json
1.01 kB
Upload folder using huggingface_hub
9 months ago
generation_config.json
136 Bytes
Upload Llama 3 license, use policy and modified generation_config
9 months ago
model.safetensors
5.74 GB
LFS
Rename gptq_model-4bit-128g.safetensors to model.safetensors
9 months ago
quantize_config.json
264 Bytes
Upload folder using huggingface_hub
9 months ago
special_tokens_map.json
301 Bytes
Upload folder using huggingface_hub
9 months ago
tokenizer.json
9.08 MB
Upload folder using huggingface_hub
9 months ago
tokenizer_config.json
50.6 kB
Upload folder using huggingface_hub
9 months ago