nougat-base-onnx / README.md
pszemraj's picture
Update README.md
c7a2514
|
raw
history blame
889 Bytes
---
license: cc-by-4.0
tags:
- nougat
- ocr
---
# nougat-base onnx
https://huggingface.co/facebook/nougat-base but exported to onnx. This is **not quantized**.
```python
from transformers import NougatProcessor
from optimum.onnxruntime import ORTModelForVision2Seq
model_name = 'pszemraj/nougat-base-onnx'
processor = NougatProcessor.from_pretrained(model_name)
model = ORTModelForVision2Seq.from_pretrained(
model_name,
provider="CPUExecutionProvider", # 'CUDAExecutionProvider' for gpu
use_merged=False,
use_io_binding=True
)
```
on colab CPU-only (_at time of writing_) you may get `CuPy` errors, to solve this uninstall it:
```sh
pip uninstall cupy-cuda11x -y
```
## how do da inference?
See [here](https://github.com/NielsRogge/Transformers-Tutorials/blob/b46d3e89e631701ef205297435064ab780c4853a/Nougat/Inference_with_Nougat_to_read_scientific_PDFs.ipynb)