Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

OnnxFloatToFloat16: Use ort float16 converter #1132

Merged
merged 4 commits into from
May 3, 2024
Merged

Conversation

jambayk
Copy link
Contributor

@jambayk jambayk commented May 1, 2024

Describe your changes

  • onnxconverter-common's float16 converter tool is not regularly maintained and it also cannot large models (>2GB). Use the float16 converted from onnxruntime which is a modified version of the previous tool and has more features.
  • Added unit test for pass.

Checklist before requesting a review

  • Add unit tests for this change.
  • Make sure all tests can pass.
  • Update documents if necessary.
  • Lint and apply fixes to your code by running lintrunner -a
  • Is this a user-facing change? If yes, give a description of this change to be included in the release notes.
  • Is this PR including examples changes? If yes, please remember to update example documentation in a follow-up PR.

(Optional) Issue link

@jambayk jambayk changed the title OnnxFloatToFloat16: use ort float16 converter, OrtTransformersOptimization: warning for phi torchscript model Use ort float16 converter, warning for phi torchscript model in transformers optimization pass May 1, 2024
@jambayk jambayk changed the title Use ort float16 converter, warning for phi torchscript model in transformers optimization pass Use ort float16 converter, Warning for phi torchscript model May 1, 2024
@jambayk jambayk marked this pull request as draft May 2, 2024 19:43
@jambayk
Copy link
Contributor Author

jambayk commented May 2, 2024

Converting to draft. Will revert warning changes.

@jambayk jambayk changed the title Use ort float16 converter, Warning for phi torchscript model OnnxFloatToFloat16: Use ort float16 converter May 2, 2024
@jambayk jambayk marked this pull request as ready for review May 2, 2024 22:36
@jambayk jambayk merged commit 3120697 into main May 3, 2024
35 checks passed
@jambayk jambayk deleted the jambayk/phi-and-fp16 branch May 3, 2024 16:23
DavitGrigoryan132 pushed a commit to DavitGrigoryan132/Olive that referenced this pull request Aug 14, 2024
## Describe your changes
- `onnxconverter-common`'s float16 converter tool is not regularly
maintained and it also cannot large models (>2GB). Use the float16
converted from onnxruntime which is a modified version of the previous
tool and has more features.
- Added unit test for pass.

## Checklist before requesting a review
- [x] Add unit tests for this change.
- [x] Make sure all tests can pass.
- [x] Update documents if necessary.
- [x] Lint and apply fixes to your code by running `lintrunner -a`
- [ ] Is this a user-facing change? If yes, give a description of this
change to be included in the release notes.
- [ ] Is this PR including examples changes? If yes, please remember to
update [example
documentation](https://github.com/microsoft/Olive/blob/main/docs/source/examples.md)
in a follow-up PR.

## (Optional) Issue link
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants