mattritchey commited on
Commit
35ce4d3
·
verified ·
1 Parent(s): d71a7c4

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +164 -0
README.md ADDED
@@ -0,0 +1,164 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ base_model: dnhkng/RYS-Medium
4
+ tags:
5
+ - llama-cpp
6
+ - gguf-my-repo
7
+ model-index:
8
+ - name: Medium
9
+ results:
10
+ - task:
11
+ type: text-generation
12
+ name: Text Generation
13
+ dataset:
14
+ name: IFEval (0-Shot)
15
+ type: HuggingFaceH4/ifeval
16
+ args:
17
+ num_few_shot: 0
18
+ metrics:
19
+ - type: inst_level_strict_acc and prompt_level_strict_acc
20
+ value: 44.06
21
+ name: strict accuracy
22
+ - type: inst_level_strict_acc and prompt_level_strict_acc
23
+ value: 44.06
24
+ name: strict accuracy
25
+ source:
26
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=dnhkng/Medium
27
+ name: Open LLM Leaderboard
28
+ - task:
29
+ type: text-generation
30
+ name: Text Generation
31
+ dataset:
32
+ name: BBH (3-Shot)
33
+ type: BBH
34
+ args:
35
+ num_few_shot: 3
36
+ metrics:
37
+ - type: acc_norm
38
+ value: 47.73
39
+ name: normalized accuracy
40
+ - type: acc_norm
41
+ value: 47.73
42
+ name: normalized accuracy
43
+ source:
44
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=dnhkng/Medium
45
+ name: Open LLM Leaderboard
46
+ - task:
47
+ type: text-generation
48
+ name: Text Generation
49
+ dataset:
50
+ name: MATH Lvl 5 (4-Shot)
51
+ type: hendrycks/competition_math
52
+ args:
53
+ num_few_shot: 4
54
+ metrics:
55
+ - type: exact_match
56
+ value: 7.78
57
+ name: exact match
58
+ - type: exact_match
59
+ value: 7.78
60
+ name: exact match
61
+ source:
62
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=dnhkng/Medium
63
+ name: Open LLM Leaderboard
64
+ - task:
65
+ type: text-generation
66
+ name: Text Generation
67
+ dataset:
68
+ name: GPQA (0-shot)
69
+ type: Idavidrein/gpqa
70
+ args:
71
+ num_few_shot: 0
72
+ metrics:
73
+ - type: acc_norm
74
+ value: 10.4
75
+ name: acc_norm
76
+ - type: acc_norm
77
+ value: 10.4
78
+ name: acc_norm
79
+ source:
80
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=dnhkng/Medium
81
+ name: Open LLM Leaderboard
82
+ - task:
83
+ type: text-generation
84
+ name: Text Generation
85
+ dataset:
86
+ name: MuSR (0-shot)
87
+ type: TAUR-Lab/MuSR
88
+ args:
89
+ num_few_shot: 0
90
+ metrics:
91
+ - type: acc_norm
92
+ value: 8.73
93
+ name: acc_norm
94
+ - type: acc_norm
95
+ value: 8.73
96
+ name: acc_norm
97
+ source:
98
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=dnhkng/Medium
99
+ name: Open LLM Leaderboard
100
+ - task:
101
+ type: text-generation
102
+ name: Text Generation
103
+ dataset:
104
+ name: MMLU-PRO (5-shot)
105
+ type: TIGER-Lab/MMLU-Pro
106
+ config: main
107
+ split: test
108
+ args:
109
+ num_few_shot: 5
110
+ metrics:
111
+ - type: acc
112
+ value: 36.96
113
+ name: accuracy
114
+ - type: acc
115
+ value: 36.96
116
+ name: accuracy
117
+ source:
118
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=dnhkng/Medium
119
+ name: Open LLM Leaderboard
120
+ ---
121
+
122
+ # mattritchey/RYS-Medium-Q4_K_M-GGUF
123
+ This model was converted to GGUF format from [`dnhkng/RYS-Medium`](https://huggingface.co/dnhkng/RYS-Medium) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
124
+ Refer to the [original model card](https://huggingface.co/dnhkng/RYS-Medium) for more details on the model.
125
+
126
+ ## Use with llama.cpp
127
+ Install llama.cpp through brew (works on Mac and Linux)
128
+
129
+ ```bash
130
+ brew install llama.cpp
131
+
132
+ ```
133
+ Invoke the llama.cpp server or the CLI.
134
+
135
+ ### CLI:
136
+ ```bash
137
+ llama-cli --hf-repo mattritchey/RYS-Medium-Q4_K_M-GGUF --hf-file rys-medium-q4_k_m.gguf -p "The meaning to life and the universe is"
138
+ ```
139
+
140
+ ### Server:
141
+ ```bash
142
+ llama-server --hf-repo mattritchey/RYS-Medium-Q4_K_M-GGUF --hf-file rys-medium-q4_k_m.gguf -c 2048
143
+ ```
144
+
145
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
146
+
147
+ Step 1: Clone llama.cpp from GitHub.
148
+ ```
149
+ git clone https://github.com/ggerganov/llama.cpp
150
+ ```
151
+
152
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
153
+ ```
154
+ cd llama.cpp && LLAMA_CURL=1 make
155
+ ```
156
+
157
+ Step 3: Run inference through the main binary.
158
+ ```
159
+ ./llama-cli --hf-repo mattritchey/RYS-Medium-Q4_K_M-GGUF --hf-file rys-medium-q4_k_m.gguf -p "The meaning to life and the universe is"
160
+ ```
161
+ or
162
+ ```
163
+ ./llama-server --hf-repo mattritchey/RYS-Medium-Q4_K_M-GGUF --hf-file rys-medium-q4_k_m.gguf -c 2048
164
+ ```