Support int8 quantization in decoder

This commit is contained in:
pkufool 2023-06-29 12:15:06 +08:00
parent 9c2172c1c4
commit 3544665ace
2 changed files with 2 additions and 2 deletions

View File

@ -757,7 +757,7 @@ def main():
quantize_dynamic(
model_input=decoder_filename,
model_output=decoder_filename_int8,
op_types_to_quantize=["MatMul"],
op_types_to_quantize=["MatMul", "Gather"],
weight_type=QuantType.QInt8,
)

View File

@ -602,7 +602,7 @@ def main():
quantize_dynamic(
model_input=decoder_filename,
model_output=decoder_filename_int8,
op_types_to_quantize=["MatMul"],
op_types_to_quantize=["MatMul", "Gather"],
weight_type=QuantType.QInt8,
)