-
Notifications
You must be signed in to change notification settings - Fork 72
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
flan-t5 support #68
Comments
can you please provide the reproducible code and output you are getting? |
does it work with python ort or are you facing the issue only with the js version of ort? |
I'm facing this error in transformer-js that is using ONNX converted model here |
Yes, I will fork the original repo and apply the changes. |
@Ki6an here my fork where you can try it This wil install the app and convert flan-t5-small to onnx:
You will find then the quantized models in the to run it
now points to http://localhost:8152/?model_id=google/flan-t5-small The tokenizer code is located here. |
I have converted google flan-t5-small using
fastT5.export_and_get_onnx_model
method with quantization enabled by defaults:getting the quantized onnx models:
Anyways when loading the model with a ONNX runtime
ort.InferenceSession
:generated tokens look strange.
Using the same process for the t5-small it works fine.
The text was updated successfully, but these errors were encountered: