[BUGFIX] Optimize preparation of self attention operators#20682
[BUGFIX] Optimize preparation of self attention operators#20682szha merged 1 commit intoapache:masterfrom
Conversation
|
Hey @bgawrych , Thanks for submitting the PR
CI supported jobs: [windows-cpu, clang, website, sanity, centos-cpu, unix-gpu, centos-gpu, unix-cpu, edge, windows-gpu, miscellaneous] Note: |
b5f2c9c to
ab9d74a
Compare
ab9d74a to
f744293
Compare
|
Is there a test that covers the modified code path? |
@szha There is a test for these operators https://github.com/apache/incubator-mxnet/blob/75e4d1d3e41cc5aef9a0141e40127a75e02ccd04/tests/python/dnnl/subgraphs/test_matmul_subgraph.py#L33 These change only removes redundant data reorder |
Description
This PR fixes problem with redundant double reorder which occurs when dealing with DNNLData