Upload folder using huggingface_hub
Browse files- README.md +9 -3
- pytorch_model.bin +2 -2
README.md
CHANGED
@@ -23,10 +23,16 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
23 |
model = AutoModelForCausalLM.from_pretrained("bonur/smollm-turkish-base")
|
24 |
tokenizer = AutoTokenizer.from_pretrained("bonur/smollm-turkish-base")
|
25 |
|
26 |
-
# Example usage
|
27 |
text = "Your prompt here"
|
28 |
-
inputs = tokenizer(text, return_tensors="pt")
|
29 |
-
outputs = model.generate(
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
30 |
result = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
31 |
print(result)
|
32 |
```
|
|
|
23 |
model = AutoModelForCausalLM.from_pretrained("bonur/smollm-turkish-base")
|
24 |
tokenizer = AutoTokenizer.from_pretrained("bonur/smollm-turkish-base")
|
25 |
|
|
|
26 |
text = "Your prompt here"
|
27 |
+
inputs = tokenizer(text, return_tensors="pt", padding=True)
|
28 |
+
outputs = model.generate(
|
29 |
+
inputs.input_ids,
|
30 |
+
attention_mask=inputs.attention_mask,
|
31 |
+
max_new_tokens=100,
|
32 |
+
do_sample=True,
|
33 |
+
temperature=0.7,
|
34 |
+
top_p=0.9
|
35 |
+
)
|
36 |
result = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
37 |
print(result)
|
38 |
```
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:049c26b844b79121ddd8379f7f69194e63f6fbf6aa007eeac0c66f17eebb8893
|
3 |
+
size 888
|