|
--- |
|
language: |
|
- en |
|
tags: |
|
- openvino |
|
--- |
|
|
|
# ibm-granite/granite-8b-code-instruct |
|
|
|
This is the [ibm-granite/granite-8b-code-instruct](https://huggingface.co/ibm-granite/granite-8b-code-instruct) model converted to [OpenVINO](https://openvino.ai) with INT8 weights compression for accelerated inference. |
|
|
|
An example of how to do inference on this model: |
|
```python |
|
from optimum.intel import OVModelForCausalLM |
|
from transformers import AutoTokenizer, pipeline |
|
|
|
# model_id should be set to either a local directory or a model available on the HuggingFace hub. |
|
model_id = "helenai/ibm-granite-granite-8b-code-instruct-ov" |
|
tokenizer = AutoTokenizer.from_pretrained(model_id) |
|
model = OVModelForCausalLM.from_pretrained(model_id) |
|
pipe = pipeline("text-generation", model=model, tokenizer=tokenizer) |
|
result = pipe("hello world") |
|
print(result) |
|
``` |
|
|
|
|