michaelfeil
commited on
Readme with Neuron
Browse files
README.md
CHANGED
@@ -135,6 +135,27 @@ Whitepaper coming soon!
|
|
135 |
|
136 |
Gradient is accelerating AI transformation across industries. https://gradient.ai/
|
137 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
138 |
## Contact Us
|
139 |
|
140 |
Drop an email to [[email protected]](mailto:[email protected])
|
|
|
135 |
|
136 |
Gradient is accelerating AI transformation across industries. https://gradient.ai/
|
137 |
|
138 |
+
## Usage with AWS Neuron
|
139 |
+
```
|
140 |
+
from transformers import AutoTokenizer
|
141 |
+
from optimum.neuron import NeuronModelForCausalLM
|
142 |
+
|
143 |
+
# Instantiate and convert to Neuron a PyTorch checkpoint
|
144 |
+
model = NeuronModelForCausalLM.from_pretrained("gradientai/v-alpha-tross")
|
145 |
+
|
146 |
+
tokenizer = AutoTokenizer.from_pretrained("gradientai/v-alpha-tross")
|
147 |
+
|
148 |
+
tokens = tokenizer("I really wish ", return_tensors="pt")
|
149 |
+
with torch.inference_mode():
|
150 |
+
sample_output = model.generate(
|
151 |
+
**tokens,
|
152 |
+
min_length=16,
|
153 |
+
max_length=32,
|
154 |
+
)
|
155 |
+
outputs = [tokenizer.decode(tok) for tok in sample_output]
|
156 |
+
print(outputs)
|
157 |
+
```
|
158 |
+
|
159 |
## Contact Us
|
160 |
|
161 |
Drop an email to [[email protected]](mailto:[email protected])
|