munish0838 commited on
Commit
3519282
1 Parent(s): 1410b67

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +92 -0
README.md ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ license: apache-2.0
5
+ language:
6
+ - en
7
+ pipeline_tag: text-generation
8
+ tags:
9
+ - chat
10
+ - qwen
11
+ - qwen2
12
+ - finetune
13
+ - chatml
14
+ - OpenHermes-2.5
15
+ - HelpSteer2
16
+ - Orca
17
+ - SlimOrca
18
+ library_name: transformers
19
+ inference: false
20
+ model_creator: MaziyarPanahi
21
+ quantized_by: MaziyarPanahi
22
+ base_model: Qwen/Qwen2-7B
23
+ model_name: calme-2.8-qwen2-7b
24
+ datasets:
25
+ - nvidia/HelpSteer2
26
+ - teknium/OpenHermes-2.5
27
+ - microsoft/orca-math-word-problems-200k
28
+ - Open-Orca/SlimOrca
29
+
30
+ ---
31
+
32
+ ![](https://cdn.discordapp.com/attachments/791342238541152306/1264099835221381251/image.png?ex=669ca436&is=669b52b6&hm=129f56187c31e1ed22cbd1bcdbc677a2baeea5090761d2f1a458c8b1ec7cca4b&)
33
+
34
+ # QuantFactory/calme-2.8-qwen2-7b-GGUF
35
+ This is quantized version of [MaziyarPanahi/calme-2.8-qwen2-7b](https://huggingface.co/MaziyarPanahi/calme-2.8-qwen2-7b) created using llama.cpp
36
+
37
+ # Original Model Card
38
+
39
+
40
+ <img src="./qwen2-fine-tunes-maziyar-panahi.webp" alt="Qwen2 fine-tune" width="500" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
41
+
42
+ # MaziyarPanahi/calme-2.8-qwen2-7b
43
+
44
+ This is a fine-tuned version of the `Qwen/Qwen2-7B` model. It aims to improve the base model across all benchmarks.
45
+
46
+ # ⚡ Quantized GGUF
47
+
48
+ All GGUF models are available here: [MaziyarPanahi/calme-2.8-qwen2-7b-GGUF](https://huggingface.co/MaziyarPanahi/calme-2.8-qwen2-7b-GGUF)
49
+
50
+ # 🏆 [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
51
+
52
+ coming soon!
53
+
54
+
55
+ # Prompt Template
56
+
57
+ This model uses `ChatML` prompt template:
58
+
59
+ ```
60
+ <|im_start|>system
61
+ {System}
62
+ <|im_end|>
63
+ <|im_start|>user
64
+ {User}
65
+ <|im_end|>
66
+ <|im_start|>assistant
67
+ {Assistant}
68
+ ````
69
+
70
+ # How to use
71
+
72
+
73
+ ```python
74
+
75
+ # Use a pipeline as a high-level helper
76
+
77
+ from transformers import pipeline
78
+
79
+ messages = [
80
+ {"role": "user", "content": "Who are you?"},
81
+ ]
82
+ pipe = pipeline("text-generation", model="MaziyarPanahi/calme-2.8-qwen2-7b")
83
+ pipe(messages)
84
+
85
+
86
+ # Load model directly
87
+
88
+ from transformers import AutoTokenizer, AutoModelForCausalLM
89
+
90
+ tokenizer = AutoTokenizer.from_pretrained("MaziyarPanahi/calme-2.8-qwen2-7b")
91
+ model = AutoModelForCausalLM.from_pretrained("MaziyarPanahi/calme-2.8-qwen2-7b")
92
+ ```