-
Notifications
You must be signed in to change notification settings - Fork 184
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Extend float16 converter api with auto_mixed_precision #543
Conversation
This pull request introduces 1 alert when merging 1fe71a6 into 92caa72 - view on LGTM.com new alerts:
|
If |
Hi @garymm, I was the original author of the We used this function a lot in our scenarios. We maintain a DL platform called AdsBrain which serves lots of users' models within Microsoft (mainly for Microsoft Ads, and also have users in MSR, Search and News, etc.). I have two concerns for deleting this function. |
My thoughts are these are apis on two different levels. Essentially the main fp16 conversion problem is that some (pt) operators don't have (fp16) implementation in onnx, thus directly converting the converted onnx subgraph to fp16 may result in under/overflow, which lead to issues such as the nan issue you observed.
@yetingqiaqia for your concern, I think |
Thanks @BowenBao. I totally agree. |
Given that, does it make sense to
? |
Created issue #544 to track next steps. |
@xadupre could you please review? |
Signed-off-by: BowenBao <bowbao@microsoft.com>
This pull request introduces 1 alert when merging 968a7f1 into 92caa72 - view on LGTM.com new alerts:
|
@xadupre could you please merge? |
@yetingqiaqia Hi Ting, I saw you shared the model in google drive, ConvNext-ML_mldelAndScript.zip. is this model > 2G? |
Thanks @BowenBao @xiaowuhu and @garymm . It turns out this |
No description provided.