set-user-agent
#11
by
jrb
- opened
- README.md +0 -5
- config.json +1 -1
README.md
CHANGED
@@ -9,7 +9,6 @@ inference:
|
|
9 |
max_new_tokens: 5
|
10 |
temperature: 1.0
|
11 |
top_k: 1
|
12 |
-
license: apache-2.0
|
13 |
language:
|
14 |
- en
|
15 |
pipeline_tag: text-generation
|
@@ -106,10 +105,6 @@ tokenizer = AutoTokenizer.from_pretrained("togethercomputer/GPT-JT-6B-v1")
|
|
106 |
model = AutoModelForCausalLM.from_pretrained("togethercomputer/GPT-JT-6B-v1")
|
107 |
```
|
108 |
|
109 |
-
# License
|
110 |
-
|
111 |
-
The weights of GPT-JT-6B-v1 are licensed under version 2.0 of the Apache License.
|
112 |
-
|
113 |
# Training Details
|
114 |
|
115 |
## UL2 Training Objective
|
|
|
9 |
max_new_tokens: 5
|
10 |
temperature: 1.0
|
11 |
top_k: 1
|
|
|
12 |
language:
|
13 |
- en
|
14 |
pipeline_tag: text-generation
|
|
|
105 |
model = AutoModelForCausalLM.from_pretrained("togethercomputer/GPT-JT-6B-v1")
|
106 |
```
|
107 |
|
|
|
|
|
|
|
|
|
108 |
# Training Details
|
109 |
|
110 |
## UL2 Training Objective
|
config.json
CHANGED
@@ -30,7 +30,7 @@
|
|
30 |
"text-generation": {
|
31 |
"do_sample": true,
|
32 |
"max_length": 50,
|
33 |
-
"temperature": 0
|
34 |
}
|
35 |
},
|
36 |
"tie_word_embeddings": false,
|
|
|
30 |
"text-generation": {
|
31 |
"do_sample": true,
|
32 |
"max_length": 50,
|
33 |
+
"temperature": 1.0
|
34 |
}
|
35 |
},
|
36 |
"tie_word_embeddings": false,
|