elysiantech commited on
Commit
064a1a7
1 Parent(s): eea595e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -17
README.md CHANGED
@@ -7,6 +7,9 @@ license_name: gemma-terms-of-use
7
  license_link: https://ai.google.dev/gemma/terms
8
  tags:
9
  - text-generation-inference
 
 
 
10
  extra_gated_heading: Access Gemma on Hugging Face
11
  extra_gated_prompt: To access Gemma on Hugging Face, you’re required to review and
12
  agree to Google’s usage license. To do this, please ensure you’re logged-in to Hugging
@@ -14,24 +17,14 @@ extra_gated_prompt: To access Gemma on Hugging Face, you’re required to review
14
  extra_gated_button_content: Acknowledge license
15
  ---
16
 
17
- GPTQ quantized version of gemma-2b model.
18
 
19
- ---
20
-
21
- [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1DTT1F5dV2wpZEU3epLTWL4WaOUjC6KAN?usp=sharing)
22
-
23
-
24
- # Gemma Model Card
25
- **Model Page**: [Gemma](https://ai.google.dev/gemma/docs)
26
-
27
- This model card corresponds to the 2B base version of the Gemma model. You can also visit the model card of the [2B base model](https://huggingface.co/google/gemma-2b)
28
-
29
- **Resources and Technical Documentation**:
30
 
31
- * [Responsible Generative AI Toolkit](https://ai.google.dev/responsible)
32
- * [Gemma on Kaggle](https://www.kaggle.com/models/google/gemma)
33
- * [Gemma on Vertex Model Garden](https://console.cloud.google.com/vertex-ai/publishers/google/model-garden/335?version=gemma-7b-gg-hf)
34
 
35
- **Terms of Use**: [Terms](https://www.kaggle.com/models/google/gemma/license/consent)
36
 
37
- **Authors**: Google
 
 
 
7
  license_link: https://ai.google.dev/gemma/terms
8
  tags:
9
  - text-generation-inference
10
+ - gemma
11
+ - gptq
12
+ - google
13
  extra_gated_heading: Access Gemma on Hugging Face
14
  extra_gated_prompt: To access Gemma on Hugging Face, you’re required to review and
15
  agree to Google’s usage license. To do this, please ensure you’re logged-in to Hugging
 
17
  extra_gated_button_content: Acknowledge license
18
  ---
19
 
20
+ [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1kFznlPlWYOrcgd7Q1NI2tYMLH_vTRuys?usp=sharing)
21
 
22
+ # elysiantech/gemma-2b-gptq-4bit
 
 
 
 
 
 
 
 
 
 
23
 
24
+ gemma-2b-gptq-4bit is a version of the [2B base model](https://huggingface.co/google/gemma-2b) model that was quantized using the GPTQ method developed by [Lin et al. (2023)](https://arxiv.org/abs/2308.07662).
 
 
25
 
26
+ Please refer to the [Original Gemma Model Card](https://ai.google.dev/gemma/docs) for details about the model preparation and training processes.
27
 
28
+ ## Dependencies
29
+ - [`auto-gptq](https://pypi.org/project/auto-gptq/0.7.1/) – [AutoGPTQ](https://github.com/PanQiWei/AutoGPTQ.git) was used to quantize the phi-3 model.
30
+ - [`vllm==0.4.2`](https://pypi.org/project/vllm/0.4.2/) – [vLLM](https://github.com/vllm-project/vllm) was used to host models for benchmarking.