mirror of
https://github.com/saymrwulf/pytorch.git
synced 2026-05-14 20:57:59 +00:00
Add support for MatMul to BatchMatMulFP16Acc{16,32}Fake Op Mapping
Test Plan: f276981395 Reviewed By: hx89 Differential Revision: D28815646 fbshipit-source-id: c16b081bf3da2b157b9d42ea67b03dae88e82c6d
This commit is contained in:
parent
599f5058cf
commit
580831bfbb
1 changed files with 2 additions and 0 deletions
|
|
@ -43,6 +43,7 @@ std::unordered_map<std::string, std::string> getFakeFp16OpMapping(
|
|||
"SparseLengthsWeightedSumFused8BitRowwiseFakeFP16NNPI"},
|
||||
{"SparseLengthsMeanFused8BitRowwise",
|
||||
"SparseLengthsMeanFused8BitRowwiseFakeFP16AccFP16"},
|
||||
{"MatMul", "BatchMatMulFP16Acc32Fake"},
|
||||
{"BatchMatMul", "BatchMatMulFP16Acc32Fake"},
|
||||
{"Sigmoid", "SigmoidFakeFp16"},
|
||||
{"SpatialBN", "SpatialBNFakeFp16NNPI"},
|
||||
|
|
@ -60,6 +61,7 @@ std::unordered_map<std::string, std::string> getFakeFp16OpMapping(
|
|||
fake_fp16_op_conversion_map["FC"] = "Fp16FCAcc16NNPI";
|
||||
fake_fp16_op_conversion_map["FbFCPacked"] = "Fp16FCAcc16NNPI";
|
||||
fake_fp16_op_conversion_map["BatchMatMul"] = "BatchMatMulFP16Acc16Fake";
|
||||
fake_fp16_op_conversion_map["MatMul"] = "BatchMatMulFP16Acc16Fake";
|
||||
}
|
||||
if (use_nnpi) {
|
||||
fake_fp16_op_conversion_map["Sigmoid"] = "SigmoidFakeFp16NNPI";
|
||||
|
|
|
|||
Loading…
Reference in a new issue