Skip to content

Conversation

@alncat
Copy link
Contributor

@alncat alncat commented Feb 4, 2021

PR types

Bug fixes

PR changes

Others

Describe

added support for fake_quantize_dequantize_abs_max op in quantization inference and compute the channel wise weight scale directly from weights

@paddle-bot-old
Copy link

paddle-bot-old bot commented Feb 4, 2021

Thanks for your contribution!
Please wait for the result of CI firstly. See Paddle CI Manual for details.

@paddle-bot-old
Copy link

paddle-bot-old bot commented Feb 4, 2021

✅ This PR's description meets the template requirements!
Please wait for other CI results.

weight_scale[j] = std::max(weight_scale[j], abs_max);
}
}
}
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

quant_axis等于其他值呢?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

这里是参考的fake_quantize_op.cc里面对于该OP的实现,里面只有两种quant_axis

@paddle-bot-old
Copy link

Sorry to inform you that aafb48e's CIs have passed for more than 7 days. To prevent PR conflicts, you need to re-run all CIs manually.

cryoco
cryoco previously approved these changes Feb 22, 2021
Copy link

@cryoco cryoco left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

Copy link
Member

@shangzhizhou shangzhizhou left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

Copy link

@cryoco cryoco left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@cryoco cryoco merged commit 5d6a8c7 into PaddlePaddle:develop Feb 23, 2021
alncat added a commit to alncat/Paddle that referenced this pull request Feb 23, 2021
PaddlePaddle#30896)

* added support for fake_quantize_dequantize_abs_max op in quantization inference pass

* remove const_cast to pass ci

* remove compare operator to pass ci-coverage

* added detailed error message for unregistered tensorrt_subgrah_pass
@alncat alncat deleted the quant_fix branch February 23, 2021 13:30
Superjomn pushed a commit that referenced this pull request Feb 24, 2021
#30896) (#31162)

* added support for fake_quantize_dequantize_abs_max op in quantization inference pass

* remove const_cast to pass ci

* remove compare operator to pass ci-coverage

* added detailed error message for unregistered tensorrt_subgrah_pass
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants