daqc
/

Text Generation
Transformers
Safetensors
Spanish
gemma
Legal
Law
Peru
Leyes Juridicas
conversational
text-generation-inference
Inference Endpoints
daqc commited on
Commit
d0ac341
1 Parent(s): 8ffb3b0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -9
README.md CHANGED
@@ -54,7 +54,7 @@ The kuntur-peru-legal-es-gemma-2b-it-merged model is a state-of-the-art language
54
  + [QLoRA Configuration 🧮](#qlora-configuration)
55
  + [Model Merging and Saving 💾](#model-merging-and-saving)
56
  * [Logging with Wandb 📊](#logging-with-wandb)
57
- * [Impacto Ambiental 🌳](#impacto-ambiental)
58
 
59
 
60
 
@@ -102,7 +102,7 @@ The dataset encompasses a wide range of topics and provisions within the Peruvia
102
  <img src="https://cdn-uploads.huggingface.co/production/uploads/64461026e1fd8d65b27e6187/m3yAx86LN-xLEZ4Mz1ALQ.png" alt="Train Graph" width="900">
103
  </p>
104
 
105
- ## Val Progress
106
  <p align="center">
107
  <img src="https://cdn-uploads.huggingface.co/production/uploads/64461026e1fd8d65b27e6187/nuk-TgiEH8IRDjmP6_luR.png" alt="Val Graph" width="900">
108
  </p>
@@ -139,7 +139,7 @@ QLoRA (Quantization LoRA) was employed to optimize the model's computational eff
139
  - **bias:** Set to "none" to exclude bias terms from adaptation, simplifying the model architecture.
140
  - **lora_dropout:** Reduced to 0.025 from the default 0.05, controlling the dropout rate during adaptation.
141
  - **task_type:** Configured as "CAUSAL_LM" to indicate the task type of the language model.
142
- -
143
  ```python
144
  config = LoraConfig(
145
  r=8,
@@ -154,18 +154,32 @@ QLoRA (Quantization LoRA) was employed to optimize the model's computational eff
154
  These configurations were crucial for optimizing the model's performance and resource utilization during training and inference, ensuring efficient deployment.
155
 
156
 
157
- ## Model Merging and Saving 💾
158
 
159
  After fine-tuning, the LoRA-adjusted weights were merged back with the base Gemma model to create the final kuntur-peru-legal-es-gemma-2b-it-merged. The model was then saved and made available through Hugging Face for easy access and further development.
160
 
161
 
162
- ## Logging with Wandb 📊
163
 
164
  During the training process, Wandb (Weights & Biases) was used for comprehensive logging and visualization of key metrics. Wandb's powerful tracking capabilities enabled real-time monitoring of training progress, evaluation metrics, and model performance. Through interactive dashboards and visualizations, Wandb facilitated deep insights into the training dynamics, allowing for efficient model optimization and debugging. This logging integration with Wandb enhances transparency, reproducibility, and collaboration among researchers and practitioners.
165
-
166
-
167
-
168
- ## Environmental impact 🌳
 
 
 
 
 
 
 
 
 
 
 
 
 
 
169
 
170
  The training of `kuntur-peru-legal-es-gemma-2b-it-merged` was conducted optimizing the computational expenditure required.
171
 
 
54
  + [QLoRA Configuration 🧮](#qlora-configuration)
55
  + [Model Merging and Saving 💾](#model-merging-and-saving)
56
  * [Logging with Wandb 📊](#logging-with-wandb)
57
+ * [Environmental impact 🌳](#environmental-impac)
58
 
59
 
60
 
 
102
  <img src="https://cdn-uploads.huggingface.co/production/uploads/64461026e1fd8d65b27e6187/m3yAx86LN-xLEZ4Mz1ALQ.png" alt="Train Graph" width="900">
103
  </p>
104
 
105
+ ## Eval Progress
106
  <p align="center">
107
  <img src="https://cdn-uploads.huggingface.co/production/uploads/64461026e1fd8d65b27e6187/nuk-TgiEH8IRDjmP6_luR.png" alt="Val Graph" width="900">
108
  </p>
 
139
  - **bias:** Set to "none" to exclude bias terms from adaptation, simplifying the model architecture.
140
  - **lora_dropout:** Reduced to 0.025 from the default 0.05, controlling the dropout rate during adaptation.
141
  - **task_type:** Configured as "CAUSAL_LM" to indicate the task type of the language model.
142
+
143
  ```python
144
  config = LoraConfig(
145
  r=8,
 
154
  These configurations were crucial for optimizing the model's performance and resource utilization during training and inference, ensuring efficient deployment.
155
 
156
 
157
+ ## Model Merging and Saving
158
 
159
  After fine-tuning, the LoRA-adjusted weights were merged back with the base Gemma model to create the final kuntur-peru-legal-es-gemma-2b-it-merged. The model was then saved and made available through Hugging Face for easy access and further development.
160
 
161
 
162
+ ## Logging with Wandb
163
 
164
  During the training process, Wandb (Weights & Biases) was used for comprehensive logging and visualization of key metrics. Wandb's powerful tracking capabilities enabled real-time monitoring of training progress, evaluation metrics, and model performance. Through interactive dashboards and visualizations, Wandb facilitated deep insights into the training dynamics, allowing for efficient model optimization and debugging. This logging integration with Wandb enhances transparency, reproducibility, and collaboration among researchers and practitioners.
165
+
166
+ - eval/loss:1.1386919021606443
167
+ - eval/runtime:44.2153
168
+ - eval/samples_per_second:8.707
169
+ - eval/steps_per_second:8.707
170
+ - train/epoch:49.62
171
+ - train/global_step:4,850
172
+ - train/grad_norm:3.5548949241638184
173
+ - train/learning_rate:0
174
+ - train/loss:0.8596
175
+ - train/total_flos:236,149,029,419,876,350
176
+ - train/train_loss:1.105836234535139
177
+ - train/train_runtime:13,237.4947
178
+ - train/train_samples_per_second:5.9
179
+ - train/train_steps_per_second:0.366
180
+
181
+
182
+ ## Environmental impact
183
 
184
  The training of `kuntur-peru-legal-es-gemma-2b-it-merged` was conducted optimizing the computational expenditure required.
185