@@ -200,7 +200,7 @@ def to_config_mapping(
200
200
self , config_list : List [BaseConfig ] = None , model_info : List [Tuple [str , str ]] = None
201
201
) -> OrderedDictType [Union [str , str ], OrderedDictType [str , BaseConfig ]]:
202
202
if not self .quant_lm_head :
203
- self .set_local (LM_HEAD_NAMES , RTNConfig (dtype = "fp32" ))
203
+ self .set_local (LM_HEAD_NAMES , RTNConfig (dtype = "fp32" , use_layer_wise = self . use_layer_wise , model_path = self . model_path ))
204
204
config_mapping = super ().to_config_mapping (config_list , model_info )
205
205
return config_mapping
206
206
@@ -363,7 +363,7 @@ def to_config_mapping(
363
363
self , config_list : List [BaseConfig ] = None , model_info : List [Tuple [str , str ]] = None
364
364
) -> OrderedDictType [Union [str , str ], OrderedDictType [str , BaseConfig ]]:
365
365
if not self .quant_lm_head :
366
- self .set_local (LM_HEAD_NAMES , GPTQConfig (dtype = "fp32" ))
366
+ self .set_local (LM_HEAD_NAMES , GPTQConfig (dtype = "fp32" , use_layer_wise = self . use_layer_wise , model_path = self . model_path ))
367
367
config_mapping = super ().to_config_mapping (config_list , model_info )
368
368
return config_mapping
369
369
@@ -385,7 +385,7 @@ def get_config_set_for_tuning(cls) -> Union[None, "GPTQConfig", List["GPTQConfig
385
385
@classmethod
386
386
def get_predefined_configs (cls ) -> Dict [torch_utils .ProcessorType , "GPTQConfig" ]:
387
387
pre_defined_configs : Dict [torch_utils .ProcessorType , GPTQConfig ] = {}
388
- pre_defined_configs [torch_utils .ProcessorType .Client ] = cls (use_layer_wise = True )
388
+ pre_defined_configs [torch_utils .ProcessorType .Client ] = cls (use_layer_wise = True )#, model_path=self.model_path)
389
389
pre_defined_configs [torch_utils .ProcessorType .Server ] = cls ()
390
390
return pre_defined_configs
391
391
@@ -508,7 +508,7 @@ def to_config_mapping(
508
508
self , config_list : List [BaseConfig ] = None , model_info : List [Tuple [str , str ]] = None
509
509
) -> OrderedDictType [Union [str , str ], OrderedDictType [str , BaseConfig ]]:
510
510
if not self .quant_lm_head :
511
- self .set_local (LM_HEAD_NAMES , AWQConfig (dtype = "fp32" ))
511
+ self .set_local (LM_HEAD_NAMES , AWQConfig (dtype = "fp32" , use_layer_wise = self . use_layer_wise , model_path = self . model_path ))
512
512
config_mapping = super ().to_config_mapping (config_list , model_info )
513
513
return config_mapping
514
514
@@ -815,7 +815,7 @@ def get_config_set_for_tuning(cls) -> Union[None, "AutoRoundConfig", List["AutoR
815
815
@classmethod
816
816
def get_predefined_configs (cls ) -> Dict [torch_utils .ProcessorType , "AutoRoundConfig" ]:
817
817
pre_defined_configs : Dict [torch_utils .ProcessorType , AutoRoundConfig ] = {}
818
- pre_defined_configs [torch_utils .ProcessorType .Client ] = cls (use_layer_wise = True )
818
+ pre_defined_configs [torch_utils .ProcessorType .Client ] = cls (use_layer_wise = True , model_path = self . model_path )
819
819
pre_defined_configs [torch_utils .ProcessorType .Server ] = cls ()
820
820
return pre_defined_configs
821
821
0 commit comments