Skip to content

Commit

Permalink
fix linting
Browse files Browse the repository at this point in the history
  • Loading branch information
fajin-corp committed Aug 14, 2024
1 parent e119ebc commit d54a2e2
Showing 1 changed file with 2 additions and 0 deletions.
2 changes: 2 additions & 0 deletions onnxruntime/python/tools/quantization/quantize.py
Original file line number Diff line number Diff line change
Expand Up @@ -688,6 +688,7 @@ def quantize_dynamic(
quantizer.quantize_model()
quantizer.model.save_model_to_file(model_output, use_external_data_format)


def quantize(
model_input: Union[str, Path, onnx.ModelProto],
model_output: Union[str, Path],
Expand Down Expand Up @@ -734,6 +735,7 @@ def quantize(
else:
# training package doesn't has quantize_matmul_4bits, avoid global import
from .matmul_4bits_quantizer import MatMul4BitsQuantizer, WeightOnlyQuantConfig

if isinstance(quant_config, WeightOnlyQuantConfig):
model = model_input if isinstance(model_input, onnx.ModelProto) else onnx.load(model_input)
quant = MatMul4BitsQuantizer(model, algo_config=quant_config)
Expand Down

0 comments on commit d54a2e2

Please sign in to comment.