s3nh commited on
Commit
592670c
·
1 Parent(s): e1d8fdb

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +78 -0
README.md ADDED
@@ -0,0 +1,78 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: openrail
3
+ pipeline_tag: text-generation
4
+ library_name: transformers
5
+ language:
6
+ - zh
7
+ - en
8
+ ---
9
+
10
+
11
+ ## Original model card
12
+
13
+ Buy me a coffee if you like this project ;)
14
+ <a href="https://www.buymeacoffee.com/s3nh"><img src="https://www.buymeacoffee.com/assets/img/guidelines/download-assets-sm-1.svg" alt=""></a>
15
+
16
+ #### Description
17
+
18
+ GGML Format model files for [This project](https://huggingface.co/Gryphe/MythoMax-L2-13b).
19
+
20
+
21
+ ### inference
22
+
23
+
24
+ ```python
25
+
26
+ import ctransformers
27
+
28
+ from ctransformers import AutoModelForCausalLM
29
+
30
+ model = AutoModelForCausalLM.from_pretrained(output_dir, ggml_file,
31
+ gpu_layers=32, model_type="llama")
32
+
33
+ manual_input: str = "Tell me about your last dream, please."
34
+
35
+
36
+ llm(manual_input,
37
+ max_new_tokens=256,
38
+ temperature=0.9,
39
+ top_p= 0.7)
40
+
41
+ ```
42
+
43
+
44
+
45
+ # Original model card
46
+
47
+
48
+
49
+ An improved, potentially even perfected variant of MythoMix, my [MythoLogic-L2](https://huggingface.co/Gryphe/MythoLogic-L2-13b) and [Huginn](https://huggingface.co/The-Face-Of-Goonery/Huginn-13b-FP16) merge using a highly experimental tensor type merge technique. The main difference with MythoMix is that I allowed more of Huginn to intermingle with the single tensors located at the front and end of a model, resulting in increased coherency across the entire structure.
50
+
51
+ The script and the acccompanying templates I used to produce both can [be found here](https://github.com/Gryphe/BlockMerge_Gradient/tree/main/YAML).
52
+
53
+ This model is proficient at both roleplaying and storywriting due to its unique nature.
54
+
55
+ Quantized models are available from TheBloke: [GGML](https://huggingface.co/TheBloke/MythoMax-L2-13B-GGML) - [GPTQ](https://huggingface.co/TheBloke/MythoMax-L2-13B-GPTQ) (You're the best!)
56
+
57
+ ## Model details
58
+
59
+ The idea behind this merge is that each layer is composed of several tensors, which are in turn responsible for specific functions. Using MythoLogic-L2's robust understanding as its input and Huginn's extensive writing capability as its output seems to have resulted in a model that exceeds at both, confirming my theory. (More details to be released at a later time)
60
+
61
+ This type of merge is incapable of being illustrated, as each of its 363 tensors had an unique ratio applied to it. As with my prior merges, gradients were part of these ratios to further finetune its behaviour.
62
+
63
+ ## Prompt Format
64
+
65
+ This model primarily uses Alpaca formatting, so for optimal model performance, use:
66
+ ```
67
+ <System prompt/Character Card>
68
+
69
+ ### Instruction:
70
+ Your instruction or question here.
71
+ For roleplay purposes, I suggest the following - Write <CHAR NAME>'s next reply in a chat between <YOUR NAME> and <CHAR NAME>. Write a single reply only.
72
+
73
+ ### Response:
74
+ ```
75
+
76
+ ---
77
+ license: other
78
+ ---