File size: 708 Bytes
012496a e219749 8e03a3e bbde354 3522113 25ebbf6 8e03a3e 3522113 8e03a3e 3522113 bbde354 3522113 6822221 cae5ccc 8e03a3e 3522113 9727478 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 |
#####
## Bloom2.5B Zen ##
#####
Bloom (2.5 B) Scientific Model fine-tuned on Zen knowledge
#####
## Usage ##
#####
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("MultiTrickFox/bloom-2b5_Zen")
model = AutoModelForCausalLM.from_pretrained("MultiTrickFox/bloom-2b5_Zen")
generator = pipeline('text-generation', model=model, tokenizer=tokenizer)
inp = [ """Today""", """Yesterday""" ]
out = generator(
inp,
do_sample=True,
temperature=.7,
typical_p=.6,
#top_p=.9,
repetition_penalty=1.2,
max_new_tokens=666,
max_time=60, # seconds
)
for o in out: print(o[0]['generated_text'])
``` |