e-palmisano commited on
Commit
71f910e
1 Parent(s): 81a082b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +24 -10
README.md CHANGED
@@ -1,22 +1,36 @@
1
  ---
2
- base_model: model_2
 
 
 
3
  language:
 
4
  - en
5
- license: apache-2.0
6
- tags:
7
- - text-generation-inference
8
- - transformers
9
- - unsloth
10
- - qwen2
11
- - trl
12
  ---
13
 
 
 
 
 
 
14
  # Uploaded model
15
 
16
  - **Developed by:** e-palmisano
17
  - **License:** apache-2.0
18
- - **Finetuned from model :** model_2
 
 
 
 
 
 
 
 
 
 
 
 
19
 
20
  This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
21
 
22
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
1
  ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - gsarti/clean_mc4_it
5
+ - FreedomIntelligence/alpaca-gpt4-italian
6
  language:
7
+ - it
8
  - en
 
 
 
 
 
 
 
9
  ---
10
 
11
+
12
+ This model has been fine-tuned with the continuous pretraining mode of Unsloth on the gsarti/clean_mc4_it dataset (only 100k rows) to improve the Italian language. The second fine-tuning was performed on the instructed dataset FreedomIntelligence/alpaca-gpt4-italian.
13
+
14
+
15
+
16
  # Uploaded model
17
 
18
  - **Developed by:** e-palmisano
19
  - **License:** apache-2.0
20
+ - **Finetuned from model :** unsloth/Qwen2-1.5B-Instruct-bnb-4bit
21
+
22
+ ## Evaluation
23
+
24
+ For a detailed comparison of model performance, check out the [Leaderboard for Italian Language Models](https://huggingface.co/spaces/FinancialSupport/open_ita_llm_leaderboard).
25
+
26
+ Here's a breakdown of the performance metrics:
27
+
28
+ | Metric | hellaswag_it acc_norm | arc_it acc_norm | m_mmlu_it 5-shot acc | Average |
29
+ |:----------------------------|:----------------------|:----------------|:---------------------|:--------|
30
+ | **Accuracy Normalized** | | | | |
31
+
32
+
33
 
34
  This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
35
 
36
+ [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)