automerger commited on
Commit
e537b1e
1 Parent(s): 0525f9a

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,33 +1,22 @@
1
  ---
2
- base_model:
3
- - CorticalStack/shadow-clown-7B-dare
4
- - mayacinka/yam-jom-7B
5
- library_name: transformers
6
  tags:
7
- - mergekit
8
  - merge
9
- license: apache-2.0
 
 
 
 
10
  ---
11
- # merge
12
-
13
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
14
-
15
- ## Merge Details
16
- ### Merge Method
17
 
18
- This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [CorticalStack/shadow-clown-7B-dare](https://huggingface.co/CorticalStack/shadow-clown-7B-dare) as a base.
19
 
20
- ### Models Merged
21
-
22
- The following models were included in the merge:
23
  * [mayacinka/yam-jom-7B](https://huggingface.co/mayacinka/yam-jom-7B)
24
 
25
- ### Configuration
26
-
27
- The following YAML configuration was used to produce this model:
28
 
29
  ```yaml
30
-
31
  models:
32
  - model: CorticalStack/shadow-clown-7B-dare
33
  # No parameters necessary for base model
@@ -41,5 +30,29 @@ parameters:
41
  int8_mask: true
42
  dtype: bfloat16
43
  random_seed: 0
44
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
  ```
 
1
  ---
2
+ license: apache-2.0
 
 
 
3
  tags:
 
4
  - merge
5
+ - mergekit
6
+ - lazymergekit
7
+ - automerger
8
+ base_model:
9
+ - mayacinka/yam-jom-7B
10
  ---
 
 
 
 
 
 
11
 
12
+ # ShadowYam-7B
13
 
14
+ ShadowYam-7B is an automated merge created by [Maxime Labonne](https://huggingface.co/mlabonne) using the following configuration.
 
 
15
  * [mayacinka/yam-jom-7B](https://huggingface.co/mayacinka/yam-jom-7B)
16
 
17
+ ## 🧩 Configuration
 
 
18
 
19
  ```yaml
 
20
  models:
21
  - model: CorticalStack/shadow-clown-7B-dare
22
  # No parameters necessary for base model
 
30
  int8_mask: true
31
  dtype: bfloat16
32
  random_seed: 0
33
+ ```
34
+
35
+ ## 💻 Usage
36
+
37
+ ```python
38
+ !pip install -qU transformers accelerate
39
+
40
+ from transformers import AutoTokenizer
41
+ import transformers
42
+ import torch
43
+
44
+ model = "automerger/ShadowYam-7B"
45
+ messages = [{"role": "user", "content": "What is a large language model?"}]
46
+
47
+ tokenizer = AutoTokenizer.from_pretrained(model)
48
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
49
+ pipeline = transformers.pipeline(
50
+ "text-generation",
51
+ model=model,
52
+ torch_dtype=torch.float16,
53
+ device_map="auto",
54
+ )
55
+
56
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
57
+ print(outputs[0]["generated_text"])
58
  ```
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4cd32ce0272918712ae50d11de9ae4b538ed474f39887473278a7eeb55342b0c
3
  size 9825524456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16c50b3f4b60c1ea1be5af26a497630925c92db36aa55705e335066f0768bdfc
3
  size 9825524456
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:96777e741a951497b01ffdd85d759c0e304d7b66ca36feb257d7b9f24361ac26
3
  size 4657973592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c6fe86c555003e332ceadd06191d3b7fd6ed7016722eef2779019894eac79d0
3
  size 4657973592