gemma_data_science / README.md
soufyane's picture
Update README.md
d283b91 verified
|
raw
history blame
2.25 kB
---
library_name: keras-nlp
pipeline_tag: text-generation
language:
- en
tags:
- gemma2b
- gemma
- google
- gemini
- gemma data science
- gemma 2b data science
- data science model
datasets:
- soufyane/DATA_SCIENCE_QA
---
This is a [`Gemma` model](https://keras.io/api/keras_nlp/models/gemma) uploaded using the KerasNLP library and can be used with JAX, TensorFlow, and PyTorch backends.
This model is related to a `CausalLM` task.
Model config:
* **name:** gemma_backbone
* **trainable:** True
* **vocabulary_size:** 256000
* **num_layers:** 18
* **num_query_heads:** 8
* **num_key_value_heads:** 1
* **hidden_dim:** 2048
* **intermediate_dim:** 32768
* **head_dim:** 256
* **layer_norm_epsilon:** 1e-06
* **dropout:** 0
* **Model Details:**
* **Architecture:** Gemma 2b is based on a deep neural network architecture, utilizing state-of-the-art techniques in natural language processing and machine learning.
* **Fine-tuning Framework:** Gemma 2b was fine-tuned using the Keras NLP framework, which provides powerful tools for building and training neural network models specifically tailored for natural language processing tasks.
* **Training Data:** Gemma 2b was fine-tuned on a diverse set of data science datasets. https://huggingface.co/datasets/soufyane/DATA_SCIENCE_QA
* **Preprocessing:** The model incorporates standard preprocessing techniques including tokenization, normalization, and feature scaling to handle input data effectively.
**use it on kaggle:**
I recommend to use the model on kaggle(free GPU use P100) for fast responses here's the link to my notebook:
https://www.kaggle.com/code/sufyen/gemma-2b-data-science-from-hugging-face
**how to use:**
```python
#install the necessery PKGs
!pip install -q -U keras-nlp
!pip install -q -U keras>=3
import keras_nlp
from keras_nlp.models import GemmaCausalLM
import warnings
warnings.filterwarnings('ignore')
import os
#set the envirenment
os.environ["KERAS_BACKEND"] = "jax" # Or "torch" or "tensorflow".
os.environ["XLA_PYTHON_CLIENT_MEM_FRACTION"]="1.00"
#load the model from HF
model = keras_nlp.models.CausalLM.from_preset(f"hf://soufyane/gemma_data_science")
while True:
x = input("enter your question: ")
print(model.generate(f"question: {x}", max_length=256))
```