You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I finetuned gliner small v2.1 model and created onnx version of the same model using the convert_to_onnx.ipynb exmple code.
When I compared the inference time of both models, the onnx version took 50% more time.
This is how I'm loading the model:
model = GLiNER.from_pretrained(model_path, load_onnx_model=True, load_tokenizer=True)
The text was updated successfully, but these errors were encountered:
I finetuned gliner small v2.1 model and created onnx version of the same model using the convert_to_onnx.ipynb exmple code.
When I compared the inference time of both models, the onnx version took 50% more time.
This is how I'm loading the model:
model = GLiNER.from_pretrained(model_path, load_onnx_model=True, load_tokenizer=True)
The text was updated successfully, but these errors were encountered: