Update README.md
Browse files
README.md
CHANGED
@@ -13,6 +13,16 @@ A continued pretrained version of unsloth/Qwen2.5-7B model using unsloth's low r
|
|
13 |
|
14 |
For pretraining, posts from [SubMaroon/DTF_comments_Responses_Counts](https://huggingface.co/datasets/SubMaroon/DTF_Comments_Responses_Counts) were selected, deduplicated by simple `df.unique` and filtered by length of 1000 < x < 128000 tokens.
|
15 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
LoRA hyperparameters:
|
17 |
|
18 |
```
|
|
|
13 |
|
14 |
For pretraining, posts from [SubMaroon/DTF_comments_Responses_Counts](https://huggingface.co/datasets/SubMaroon/DTF_Comments_Responses_Counts) were selected, deduplicated by simple `df.unique` and filtered by length of 1000 < x < 128000 tokens.
|
15 |
|
16 |
+
The model is trained for text completion and not optimized for conversational usage. If you want to use it anyway, here is usage example:
|
17 |
+
|
18 |
+
```
|
19 |
+
import transformers
|
20 |
+
|
21 |
+
pipe = transformers.pipeline('text-generation', model='./Qwen2.5-7B-DTF-pretrain/', device=0)
|
22 |
+
print(pipe('Кто такой карасик?', max_new_tokens=45))
|
23 |
+
# >>> [{'generated_text': 'Кто такой карасик? Это неизвестный, но очень опасный и сильный враг, который может убить любого героя. Он может быть в любой момент, в любом месте, и его невозможно предугадать.'}]
|
24 |
+
```
|
25 |
+
|
26 |
LoRA hyperparameters:
|
27 |
|
28 |
```
|