@@ -387,7 +387,6 @@ class AWQConfig(BaseConfig):
387
387
"use_full_range" ,
388
388
"use_mse_search" ,
389
389
"use_layer_wise" ,
390
- "export_compressed_model" ,
391
390
"use_double_quant" ,
392
391
"double_quant_dtype" ,
393
392
"double_quant_bits" ,
@@ -410,7 +409,6 @@ def __init__(
410
409
use_full_range : bool = False ,
411
410
use_mse_search : bool = False ,
412
411
use_layer_wise : bool = False ,
413
- export_compressed_model : bool = False ,
414
412
# double quant
415
413
use_double_quant : bool = False ,
416
414
double_quant_dtype : str = "int" ,
@@ -434,7 +432,6 @@ def __init__(
434
432
use_full_range (bool): Enables full range for activations, default is False.
435
433
use_mse_search (bool): Enables mean squared error (MSE) search, default is False.
436
434
use_layer_wise (bool): Enables quantize model per layer. Defaults to False.
437
- export_compressed_model (bool): Enables return model in int format or not. Defaults to False.
438
435
use_double_quant (bool): Enables double quantization, default is False.
439
436
double_quant_dtype (str): Data type for double_quant scale, default is "int".
440
437
double_quant_bits (int): Number of bits used to represent double_quant scale, default is 4.
@@ -454,7 +451,6 @@ def __init__(
454
451
self .use_full_range = use_full_range
455
452
self .use_mse_search = use_mse_search
456
453
self .use_layer_wise = use_layer_wise
457
- self .export_compressed_model = export_compressed_model
458
454
# double quant
459
455
self .use_double_quant = use_double_quant
460
456
self .double_quant_bits = double_quant_bits
0 commit comments