mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-08-08 17:42:21 +00:00
Support int8 quantization in decoder (#1152)
This commit is contained in:
parent
9c2172c1c4
commit
db71b03026
@ -757,7 +757,7 @@ def main():
|
|||||||
quantize_dynamic(
|
quantize_dynamic(
|
||||||
model_input=decoder_filename,
|
model_input=decoder_filename,
|
||||||
model_output=decoder_filename_int8,
|
model_output=decoder_filename_int8,
|
||||||
op_types_to_quantize=["MatMul"],
|
op_types_to_quantize=["MatMul", "Gather"],
|
||||||
weight_type=QuantType.QInt8,
|
weight_type=QuantType.QInt8,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -602,7 +602,7 @@ def main():
|
|||||||
quantize_dynamic(
|
quantize_dynamic(
|
||||||
model_input=decoder_filename,
|
model_input=decoder_filename,
|
||||||
model_output=decoder_filename_int8,
|
model_output=decoder_filename_int8,
|
||||||
op_types_to_quantize=["MatMul"],
|
op_types_to_quantize=["MatMul", "Gather"],
|
||||||
weight_type=QuantType.QInt8,
|
weight_type=QuantType.QInt8,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user