Sarah Ciston
commited on
Commit
·
630ac98
1
Parent(s):
87f1324
no config
Browse files
sketch.js
CHANGED
@@ -180,10 +180,12 @@ async function runModel(PREPROMPT, PROMPT){
|
|
180 |
// , 'meta-llama/Meta-Llama-3-70B-Instruct'
|
181 |
// , 'openai-community/gpt2'
|
182 |
|
183 |
-
out = await pipe((PREPROMPT, PROMPT), {
|
184 |
-
|
185 |
-
|
186 |
-
})
|
|
|
|
|
187 |
|
188 |
console.log(out)
|
189 |
|
@@ -213,7 +215,7 @@ async function runModel(PREPROMPT, PROMPT){
|
|
213 |
// return modelResult
|
214 |
|
215 |
|
216 |
-
|
217 |
//inference.fill_mask({
|
218 |
// let out = await pipe(PREPROMPT + PROMPT)
|
219 |
// let out = await pipe(PREPROMPT + PROMPT, {
|
|
|
180 |
// , 'meta-llama/Meta-Llama-3-70B-Instruct'
|
181 |
// , 'openai-community/gpt2'
|
182 |
|
183 |
+
// out = await pipe((PREPROMPT, PROMPT), {
|
184 |
+
// max_tokens: 150,
|
185 |
+
// return_full_text: false
|
186 |
+
// })
|
187 |
+
|
188 |
+
out = await pipe((PREPROMPT, PROMPT))
|
189 |
|
190 |
console.log(out)
|
191 |
|
|
|
215 |
// return modelResult
|
216 |
|
217 |
|
218 |
+
|
219 |
//inference.fill_mask({
|
220 |
// let out = await pipe(PREPROMPT + PROMPT)
|
221 |
// let out = await pipe(PREPROMPT + PROMPT, {
|