This repository has been archived by the owner on Nov 17, 2023. It is now read-only.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
[Doc] Add MKL-DNN operator list #14891
[Doc] Add MKL-DNN operator list #14891
Changes from 18 commits
c202b36
6fbfe48
15446a8
8a3b9b9
70f3723
5b3f6db
58e8ac1
ce82caa
dab9ddf
2a5e2cd
708edee
fb5fcc3
4a61c8e
4c897e8
eb36414
bfc5ac0
7e53e8d
b74f3f7
d1cf743
762945b
6b01ce3
5c3d067
d9fcea4
b6de387
ad5e2c8
2d8e4d8
d7c7776
e2f0d03
bd49e9d
b783f58
0f82b3b
dd4cfa5
4d36bbf
File filter
Filter by extension
Conversations
Jump to
There are no files selected for viewing
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
all fusion passes
might not be accurate enough here.MKLDNN
backend is only applicable to the FP32 mode fusion, whileMKLDNN_POST_QUANTIZE
is applicable to the INT8 mode fusion (quantized_op +requantize/dequantize).There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Nice to know! How many subgraph backend options do we have now? What about INT8 convolution + relu fusion?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Currently, there're only two options as mentioned above.
There's no fusion pass for
INT8 conv +relu
, the current flow is doing the fusion ofFP32 conv +relu
firstly, and then quantizingFP32 sg_conv
toINT8 sg_conv
.There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you for the information. I think we need to put them into this page: https://mxnet.incubator.apache.org/versions/master/faq/env_var.html
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Let's maintain them in one spot and then reference where appropriate. Either way some mention is needed in the env_var page.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There's also other options for end-use to choose, such as:
MXNET_DISABLE_MKLDNN_CONV_OPT
to disable all MKLDNN convolution optimization pass.MXNET_DISABLE_MKLDNN_FC_OPT
to disable all MKLDNN FullyConnected optimization pass.There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
miss a pattern here:
FullyConnected (INT8) + re-quantization + de-quantization