elysiantech
commited on
Commit
•
064a1a7
1
Parent(s):
eea595e
Update README.md
Browse files
README.md
CHANGED
@@ -7,6 +7,9 @@ license_name: gemma-terms-of-use
|
|
7 |
license_link: https://ai.google.dev/gemma/terms
|
8 |
tags:
|
9 |
- text-generation-inference
|
|
|
|
|
|
|
10 |
extra_gated_heading: Access Gemma on Hugging Face
|
11 |
extra_gated_prompt: To access Gemma on Hugging Face, you’re required to review and
|
12 |
agree to Google’s usage license. To do this, please ensure you’re logged-in to Hugging
|
@@ -14,24 +17,14 @@ extra_gated_prompt: To access Gemma on Hugging Face, you’re required to review
|
|
14 |
extra_gated_button_content: Acknowledge license
|
15 |
---
|
16 |
|
17 |
-
|
18 |
|
19 |
-
|
20 |
-
|
21 |
-
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1DTT1F5dV2wpZEU3epLTWL4WaOUjC6KAN?usp=sharing)
|
22 |
-
|
23 |
-
|
24 |
-
# Gemma Model Card
|
25 |
-
**Model Page**: [Gemma](https://ai.google.dev/gemma/docs)
|
26 |
-
|
27 |
-
This model card corresponds to the 2B base version of the Gemma model. You can also visit the model card of the [2B base model](https://huggingface.co/google/gemma-2b)
|
28 |
-
|
29 |
-
**Resources and Technical Documentation**:
|
30 |
|
31 |
-
|
32 |
-
* [Gemma on Kaggle](https://www.kaggle.com/models/google/gemma)
|
33 |
-
* [Gemma on Vertex Model Garden](https://console.cloud.google.com/vertex-ai/publishers/google/model-garden/335?version=gemma-7b-gg-hf)
|
34 |
|
35 |
-
|
36 |
|
37 |
-
|
|
|
|
|
|
7 |
license_link: https://ai.google.dev/gemma/terms
|
8 |
tags:
|
9 |
- text-generation-inference
|
10 |
+
- gemma
|
11 |
+
- gptq
|
12 |
+
- google
|
13 |
extra_gated_heading: Access Gemma on Hugging Face
|
14 |
extra_gated_prompt: To access Gemma on Hugging Face, you’re required to review and
|
15 |
agree to Google’s usage license. To do this, please ensure you’re logged-in to Hugging
|
|
|
17 |
extra_gated_button_content: Acknowledge license
|
18 |
---
|
19 |
|
20 |
+
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1kFznlPlWYOrcgd7Q1NI2tYMLH_vTRuys?usp=sharing)
|
21 |
|
22 |
+
# elysiantech/gemma-2b-gptq-4bit
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
23 |
|
24 |
+
gemma-2b-gptq-4bit is a version of the [2B base model](https://huggingface.co/google/gemma-2b) model that was quantized using the GPTQ method developed by [Lin et al. (2023)](https://arxiv.org/abs/2308.07662).
|
|
|
|
|
25 |
|
26 |
+
Please refer to the [Original Gemma Model Card](https://ai.google.dev/gemma/docs) for details about the model preparation and training processes.
|
27 |
|
28 |
+
## Dependencies
|
29 |
+
- [`auto-gptq](https://pypi.org/project/auto-gptq/0.7.1/) – [AutoGPTQ](https://github.com/PanQiWei/AutoGPTQ.git) was used to quantize the phi-3 model.
|
30 |
+
- [`vllm==0.4.2`](https://pypi.org/project/vllm/0.4.2/) – [vLLM](https://github.com/vllm-project/vllm) was used to host models for benchmarking.
|