diff --git a/neural_compressor/torch/algorithms/weight_only/save_load.py b/neural_compressor/torch/algorithms/weight_only/save_load.py index 6a3474b2f3a..d5e6fbef235 100644 --- a/neural_compressor/torch/algorithms/weight_only/save_load.py +++ b/neural_compressor/torch/algorithms/weight_only/save_load.py @@ -230,6 +230,7 @@ def load_hf_format_woq_model(self): ): # # pragma: no cover # load autoround format quantized model from auto_round import AutoRoundConfig + hf_kargs = {} pretrain_args = ["trust_remote_code", "_attn_implementation", "device_map", "torch_dtype"] for item in pretrain_args: @@ -463,7 +464,7 @@ def _get_model_class_and_config(self): if model_cls: model_class = model_cls else: - logger.info(f"Could't find model class.") + logger.info("Couldn't find model class.") return model_class, config def _get_loaded_state_dict(self, config): @@ -921,4 +922,3 @@ def _use_hpu_module(self): # pragma: no cover if os.path.exists(os.path.join(self._model_local_dir, HPU_WEIGHT_NAME)): return True return False -