@@ -706,7 +706,7 @@ def autoround_quantize(
706
706
dynamic_max_gap : int = - 1 ,
707
707
data_type : str = "int" , ##only support int for now
708
708
scale_dtype : str = "fp16" ,
709
- quant_block_list : list = None ,
709
+ to_quant_block_names : list = None ,
710
710
act_bits : int = 32 ,
711
711
act_group_size : int = None ,
712
712
act_sym : bool = None ,
@@ -761,7 +761,7 @@ def autoround_quantize(
761
761
data_type (str): The data type to be used (default is "int").
762
762
scale_dtype (str): The data type of quantization scale to be used (default is "float32"), different kernels
763
763
have different choices.
764
- quant_block_list (list): A list whose elements are list of block's layer names to be quantized.
764
+ to_quant_block_names (list): A list whose elements are list of block's layer names to be quantized.
765
765
act_bits (int): Number of bits for activation quantization. Default is 32.
766
766
act_group_size (int): Group size for activation quantization. Default is None.
767
767
act_sym (bool): Whether to use symmetric activation quantization. Default is None.
@@ -800,7 +800,7 @@ def autoround_quantize(
800
800
dynamic_max_gap = dynamic_max_gap ,
801
801
data_type = data_type , ## only support data_type
802
802
scale_dtype = scale_dtype ,
803
- quant_block_list = quant_block_list ,
803
+ to_quant_block_names = to_quant_block_names ,
804
804
act_bits = act_bits ,
805
805
act_group_size = act_group_size ,
806
806
act_sym = act_sym ,
0 commit comments