Quantization
#16 opened about 2 months ago
by
jntjdbhvebjynfbjdv
Add exported onnx model 'model_qint8_arm64.onnx'
#15 opened about 2 months ago
by
tonatiuh-synopsys
Matryoshka embeddings
1
#14 opened 4 months ago
by
ghidav

model doesnt seem to support device_map="auto" for multi GPU .
2
#11 opened 8 months ago
by
pulkitmehtametacube
How to change the batch size? or any tips to make the model faster?
1
#7 opened 10 months ago
by
sdvfsfwfgfwf
Help with error message "Could not locate the configuration.py inside Alibaba-NLP/new-impl."
1
#6 opened 10 months ago
by
azhang01