-
Notifications
You must be signed in to change notification settings - Fork 49
New problem configs from onnxruntime examples collected on MI350 #2001
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: develop
Are you sure you want to change the base?
Conversation
-t f32 -out_datatype f32 -transA false -transB false -g 1 -m 1 -n 4096 -k 4096 | ||
-t f32 -out_datatype f32 -transA false -transB false -g 1 -m 1 -n 32000 -k 4096 | ||
#inception_v3_int8_bs16.onnx | ||
conv -F 1 -f GNC01 -I NGC01 -O NGC01 -n 16 -c 3 -H 299 -W 299 -k 32 -y 3 -x 3 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
convolution should be int8, right? this should start by convint8 instead of conv
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This was captured from the script that we have. I will double check it.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I have double checked, that is what the model produces. This is from one of the onnxruntime-inference-examples
, it seems the quantization is done at runtime for this and that is why we don't see convint8 for this. The int8
label is used because the model comes from migx_onnxrt_inception_v3_int8_benchmarks
in DLM.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
what does it mean that the quantization is done at runtime?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There is a calibration table that is used for int8 inference. I have shared more details in a Teams message with you regarding how its being done.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thanks, I've answered in teams as well. Something seems wrong, if we calibrate for int8, why don't we do inference in int8?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@dhernandez0
Problem configs are collected from inceptionv3 model that is not quantized for int8. I have updated the label name to reflect that.
…snet model configs" This reverts commit 171b6a2.
Motivation
Doing exhaustive tune for the models improves performance on MI350. We can add these models to the quick tune list to improve out of the box performance for MI350.
Technical Details
Add problem configs for models run using onnxruntime examples on MI350.
Test Plan
Test Result
Submission Checklist