nur-dev commited on
Commit
79b3cd2
1 Parent(s): a1aef4b

chore: set author ref

Browse files
Files changed (1) hide show
  1. README.md +7 -1
README.md CHANGED
@@ -34,4 +34,10 @@ for t in predicted:
34
  The model was trained using two NVIDIA A100 GPUs on over 5.3 million examples from the "kz-transformers/multidomain-kazakh-dataset." We conducted training across 10 epochs, handling large batches of data efficiently through gradient accumulation. The learning setup included a slow build-up in the learning rate to maximize learning stability and was optimized over 208,100 steps, focusing on improving the model’s ability to understand and generate the Kazakh language.
35
 
36
  ## Limitations and Bias
37
- As with any language model, roberta-kaz-large may inherently learn biases present in the training data. Users should be cautious and evaluate the model in diverse contexts to ensure it performs as expected, especially in sensitive applications.
 
 
 
 
 
 
 
34
  The model was trained using two NVIDIA A100 GPUs on over 5.3 million examples from the "kz-transformers/multidomain-kazakh-dataset." We conducted training across 10 epochs, handling large batches of data efficiently through gradient accumulation. The learning setup included a slow build-up in the learning rate to maximize learning stability and was optimized over 208,100 steps, focusing on improving the model’s ability to understand and generate the Kazakh language.
35
 
36
  ## Limitations and Bias
37
+ As with any language model, roberta-kaz-large may inherently learn biases present in the training data. Users should be cautious and evaluate the model in diverse contexts to ensure it performs as expected, especially in sensitive applications.
38
+
39
+ ## Model Authors
40
+
41
+ **Name:** Kadyrbek Nurgali
42
+ - **Email:** [email protected]
43
+ - **LinkedIn:** [Kadyrbek Nurgali](https://www.linkedin.com/in/nurgali-kadyrbek-504260231/)