Skip to content

Commit

Permalink
Fix causallm weights compression via quantizer (#484)
Browse files Browse the repository at this point in the history
  • Loading branch information
eaidova authored Dec 6, 2023
1 parent bddeacd commit 8e1ad36
Showing 1 changed file with 2 additions and 1 deletion.
3 changes: 2 additions & 1 deletion optimum/intel/openvino/quantization.py
Original file line number Diff line number Diff line change
Expand Up @@ -260,7 +260,8 @@ def _quantize_ovcausallm(
save_directory.mkdir(parents=True, exist_ok=True)

if weights_only:
self.model.model = nncf.compress_weights(self.model.model)
model = nncf.compress_weights(self.model._original_model)
self.model.model = model
self.model.save_pretrained(save_directory)
return

Expand Down

0 comments on commit 8e1ad36

Please sign in to comment.