itlwas commited on
Commit
5e321d2
1 Parent(s): 1196255

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +99 -0
README.md ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: other
5
+ tags:
6
+ - axolotl
7
+ - generated_from_trainer
8
+ - phi
9
+ - phi2
10
+ - einstein
11
+ - instruct
12
+ - finetune
13
+ - chatml
14
+ - gpt4
15
+ - synthetic data
16
+ - science
17
+ - physics
18
+ - chemistry
19
+ - biology
20
+ - math
21
+ - llama-cpp
22
+ - gguf-my-repo
23
+ base_model: Weyaxi/Einstein-v4-Qwen-1.5-32B
24
+ datasets:
25
+ - allenai/ai2_arc
26
+ - camel-ai/physics
27
+ - camel-ai/chemistry
28
+ - camel-ai/biology
29
+ - camel-ai/math
30
+ - metaeval/reclor
31
+ - openbookqa
32
+ - mandyyyyii/scibench
33
+ - derek-thomas/ScienceQA
34
+ - TIGER-Lab/ScienceEval
35
+ - jondurbin/airoboros-3.2
36
+ - LDJnr/Capybara
37
+ - Cot-Alpaca-GPT4-From-OpenHermes-2.5
38
+ - STEM-AI-mtl/Electrical-engineering
39
+ - knowrohit07/saraswati-stem
40
+ - sablo/oasst2_curated
41
+ - glaiveai/glaive-code-assistant
42
+ - lmsys/lmsys-chat-1m
43
+ - TIGER-Lab/MathInstruct
44
+ - bigbio/med_qa
45
+ - meta-math/MetaMathQA-40K
46
+ - openbookqa
47
+ - piqa
48
+ - metaeval/reclor
49
+ - derek-thomas/ScienceQA
50
+ - scibench
51
+ - sciq
52
+ - Open-Orca/SlimOrca
53
+ - migtissera/Synthia-v1.3
54
+ - TIGER-Lab/ScienceEval
55
+ ---
56
+
57
+ # AIronMind/Einstein-v4-Qwen-1.5-32B-Q4_K_M-GGUF
58
+ This model was converted to GGUF format from [`Weyaxi/Einstein-v4-Qwen-1.5-32B`](https://huggingface.co/Weyaxi/Einstein-v4-Qwen-1.5-32B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
59
+ Refer to the [original model card](https://huggingface.co/Weyaxi/Einstein-v4-Qwen-1.5-32B) for more details on the model.
60
+
61
+ ## Use with llama.cpp
62
+ Install llama.cpp through brew (works on Mac and Linux)
63
+
64
+ ```bash
65
+ brew install llama.cpp
66
+
67
+ ```
68
+ Invoke the llama.cpp server or the CLI.
69
+
70
+ ### CLI:
71
+ ```bash
72
+ llama-cli --hf-repo AIronMind/Einstein-v4-Qwen-1.5-32B-Q4_K_M-GGUF --hf-file einstein-v4-qwen-1.5-32b-q4_k_m.gguf -p "The meaning to life and the universe is"
73
+ ```
74
+
75
+ ### Server:
76
+ ```bash
77
+ llama-server --hf-repo AIronMind/Einstein-v4-Qwen-1.5-32B-Q4_K_M-GGUF --hf-file einstein-v4-qwen-1.5-32b-q4_k_m.gguf -c 2048
78
+ ```
79
+
80
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
81
+
82
+ Step 1: Clone llama.cpp from GitHub.
83
+ ```
84
+ git clone https://github.com/ggerganov/llama.cpp
85
+ ```
86
+
87
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
88
+ ```
89
+ cd llama.cpp && LLAMA_CURL=1 make
90
+ ```
91
+
92
+ Step 3: Run inference through the main binary.
93
+ ```
94
+ ./llama-cli --hf-repo AIronMind/Einstein-v4-Qwen-1.5-32B-Q4_K_M-GGUF --hf-file einstein-v4-qwen-1.5-32b-q4_k_m.gguf -p "The meaning to life and the universe is"
95
+ ```
96
+ or
97
+ ```
98
+ ./llama-server --hf-repo AIronMind/Einstein-v4-Qwen-1.5-32B-Q4_K_M-GGUF --hf-file einstein-v4-qwen-1.5-32b-q4_k_m.gguf -c 2048
99
+ ```