Update README.md
Browse files
README.md
CHANGED
@@ -54,7 +54,7 @@ The kuntur-peru-legal-es-gemma-2b-it-merged model is a state-of-the-art language
|
|
54 |
+ [QLoRA Configuration 🧮](#qlora-configuration)
|
55 |
+ [Model Merging and Saving 💾](#model-merging-and-saving)
|
56 |
* [Logging with Wandb 📊](#logging-with-wandb)
|
57 |
-
* [
|
58 |
|
59 |
|
60 |
|
@@ -102,7 +102,7 @@ The dataset encompasses a wide range of topics and provisions within the Peruvia
|
|
102 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/64461026e1fd8d65b27e6187/m3yAx86LN-xLEZ4Mz1ALQ.png" alt="Train Graph" width="900">
|
103 |
</p>
|
104 |
|
105 |
-
##
|
106 |
<p align="center">
|
107 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/64461026e1fd8d65b27e6187/nuk-TgiEH8IRDjmP6_luR.png" alt="Val Graph" width="900">
|
108 |
</p>
|
@@ -139,7 +139,7 @@ QLoRA (Quantization LoRA) was employed to optimize the model's computational eff
|
|
139 |
- **bias:** Set to "none" to exclude bias terms from adaptation, simplifying the model architecture.
|
140 |
- **lora_dropout:** Reduced to 0.025 from the default 0.05, controlling the dropout rate during adaptation.
|
141 |
- **task_type:** Configured as "CAUSAL_LM" to indicate the task type of the language model.
|
142 |
-
|
143 |
```python
|
144 |
config = LoraConfig(
|
145 |
r=8,
|
@@ -154,18 +154,32 @@ QLoRA (Quantization LoRA) was employed to optimize the model's computational eff
|
|
154 |
These configurations were crucial for optimizing the model's performance and resource utilization during training and inference, ensuring efficient deployment.
|
155 |
|
156 |
|
157 |
-
## Model Merging and Saving
|
158 |
|
159 |
After fine-tuning, the LoRA-adjusted weights were merged back with the base Gemma model to create the final kuntur-peru-legal-es-gemma-2b-it-merged. The model was then saved and made available through Hugging Face for easy access and further development.
|
160 |
|
161 |
|
162 |
-
## Logging with Wandb
|
163 |
|
164 |
During the training process, Wandb (Weights & Biases) was used for comprehensive logging and visualization of key metrics. Wandb's powerful tracking capabilities enabled real-time monitoring of training progress, evaluation metrics, and model performance. Through interactive dashboards and visualizations, Wandb facilitated deep insights into the training dynamics, allowing for efficient model optimization and debugging. This logging integration with Wandb enhances transparency, reproducibility, and collaboration among researchers and practitioners.
|
165 |
-
|
166 |
-
|
167 |
-
|
168 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
169 |
|
170 |
The training of `kuntur-peru-legal-es-gemma-2b-it-merged` was conducted optimizing the computational expenditure required.
|
171 |
|
|
|
54 |
+ [QLoRA Configuration 🧮](#qlora-configuration)
|
55 |
+ [Model Merging and Saving 💾](#model-merging-and-saving)
|
56 |
* [Logging with Wandb 📊](#logging-with-wandb)
|
57 |
+
* [Environmental impact 🌳](#environmental-impac)
|
58 |
|
59 |
|
60 |
|
|
|
102 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/64461026e1fd8d65b27e6187/m3yAx86LN-xLEZ4Mz1ALQ.png" alt="Train Graph" width="900">
|
103 |
</p>
|
104 |
|
105 |
+
## Eval Progress
|
106 |
<p align="center">
|
107 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/64461026e1fd8d65b27e6187/nuk-TgiEH8IRDjmP6_luR.png" alt="Val Graph" width="900">
|
108 |
</p>
|
|
|
139 |
- **bias:** Set to "none" to exclude bias terms from adaptation, simplifying the model architecture.
|
140 |
- **lora_dropout:** Reduced to 0.025 from the default 0.05, controlling the dropout rate during adaptation.
|
141 |
- **task_type:** Configured as "CAUSAL_LM" to indicate the task type of the language model.
|
142 |
+
|
143 |
```python
|
144 |
config = LoraConfig(
|
145 |
r=8,
|
|
|
154 |
These configurations were crucial for optimizing the model's performance and resource utilization during training and inference, ensuring efficient deployment.
|
155 |
|
156 |
|
157 |
+
## Model Merging and Saving
|
158 |
|
159 |
After fine-tuning, the LoRA-adjusted weights were merged back with the base Gemma model to create the final kuntur-peru-legal-es-gemma-2b-it-merged. The model was then saved and made available through Hugging Face for easy access and further development.
|
160 |
|
161 |
|
162 |
+
## Logging with Wandb
|
163 |
|
164 |
During the training process, Wandb (Weights & Biases) was used for comprehensive logging and visualization of key metrics. Wandb's powerful tracking capabilities enabled real-time monitoring of training progress, evaluation metrics, and model performance. Through interactive dashboards and visualizations, Wandb facilitated deep insights into the training dynamics, allowing for efficient model optimization and debugging. This logging integration with Wandb enhances transparency, reproducibility, and collaboration among researchers and practitioners.
|
165 |
+
|
166 |
+
- eval/loss:1.1386919021606443
|
167 |
+
- eval/runtime:44.2153
|
168 |
+
- eval/samples_per_second:8.707
|
169 |
+
- eval/steps_per_second:8.707
|
170 |
+
- train/epoch:49.62
|
171 |
+
- train/global_step:4,850
|
172 |
+
- train/grad_norm:3.5548949241638184
|
173 |
+
- train/learning_rate:0
|
174 |
+
- train/loss:0.8596
|
175 |
+
- train/total_flos:236,149,029,419,876,350
|
176 |
+
- train/train_loss:1.105836234535139
|
177 |
+
- train/train_runtime:13,237.4947
|
178 |
+
- train/train_samples_per_second:5.9
|
179 |
+
- train/train_steps_per_second:0.366
|
180 |
+
|
181 |
+
|
182 |
+
## Environmental impact
|
183 |
|
184 |
The training of `kuntur-peru-legal-es-gemma-2b-it-merged` was conducted optimizing the computational expenditure required.
|
185 |
|