Skip to content

Commit a0dee94

Browse files
authored
Remove export_compressed_model in AWQConfig (#1831)
1 parent 2c3556d commit a0dee94

File tree

2 files changed

+0
-7
lines changed

2 files changed

+0
-7
lines changed

neural_compressor/torch/quantization/algorithm_entry.py

-3
Original file line numberDiff line numberDiff line change
@@ -328,7 +328,6 @@ def awq_quantize_entry(
328328
"use_full_range": op_config.use_full_range,
329329
"use_mse_search": op_config.use_mse_search,
330330
"use_layer_wise": op_config.use_layer_wise,
331-
"export_compressed_model": op_config.export_compressed_model,
332331
"use_double_quant": op_config.use_double_quant,
333332
"double_quant_dtype": op_config.double_quant_dtype,
334333
"double_quant_bits": op_config.double_quant_bits,
@@ -338,7 +337,6 @@ def awq_quantize_entry(
338337
use_auto_scale = op_config.use_auto_scale
339338
use_mse_search = op_config.use_auto_clip # for awq clip
340339
folding = op_config.folding
341-
return_int = op_config.export_compressed_model
342340
use_full_range = op_config.use_full_range
343341

344342
run_fn = kwargs.get("run_fn", None)
@@ -357,7 +355,6 @@ def awq_quantize_entry(
357355
use_auto_scale=use_auto_scale,
358356
use_mse_search=use_mse_search,
359357
folding=folding,
360-
return_int=return_int,
361358
use_full_range=use_full_range,
362359
)
363360

neural_compressor/torch/quantization/config.py

-4
Original file line numberDiff line numberDiff line change
@@ -387,7 +387,6 @@ class AWQConfig(BaseConfig):
387387
"use_full_range",
388388
"use_mse_search",
389389
"use_layer_wise",
390-
"export_compressed_model",
391390
"use_double_quant",
392391
"double_quant_dtype",
393392
"double_quant_bits",
@@ -410,7 +409,6 @@ def __init__(
410409
use_full_range: bool = False,
411410
use_mse_search: bool = False,
412411
use_layer_wise: bool = False,
413-
export_compressed_model: bool = False,
414412
# double quant
415413
use_double_quant: bool = False,
416414
double_quant_dtype: str = "int",
@@ -434,7 +432,6 @@ def __init__(
434432
use_full_range (bool): Enables full range for activations, default is False.
435433
use_mse_search (bool): Enables mean squared error (MSE) search, default is False.
436434
use_layer_wise (bool): Enables quantize model per layer. Defaults to False.
437-
export_compressed_model (bool): Enables return model in int format or not. Defaults to False.
438435
use_double_quant (bool): Enables double quantization, default is False.
439436
double_quant_dtype (str): Data type for double_quant scale, default is "int".
440437
double_quant_bits (int): Number of bits used to represent double_quant scale, default is 4.
@@ -454,7 +451,6 @@ def __init__(
454451
self.use_full_range = use_full_range
455452
self.use_mse_search = use_mse_search
456453
self.use_layer_wise = use_layer_wise
457-
self.export_compressed_model = export_compressed_model
458454
# double quant
459455
self.use_double_quant = use_double_quant
460456
self.double_quant_bits = double_quant_bits

0 commit comments

Comments
 (0)