Skip to content

Commit

Permalink
style
Browse files Browse the repository at this point in the history
  • Loading branch information
dsikka committed Jan 16, 2025
1 parent 6e20e95 commit 1f6ed71
Showing 1 changed file with 0 additions and 2 deletions.
2 changes: 0 additions & 2 deletions src/transformers/quantizers/quantizer_compressed_tensors.py
Original file line number Diff line number Diff line change
Expand Up @@ -55,7 +55,6 @@ def __init__(self, quantization_config: CompressedTensorsConfig, **kwargs):
self.weight_norm = torch.nn.utils.parametrizations.weight_norm
delattr(torch.nn.utils.parametrizations, "weight_norm")


def validate_environment(self, *args, **kwargs):
if not is_compressed_tensors_available():
raise ImportError(
Expand Down Expand Up @@ -86,7 +85,6 @@ def _process_model_before_weight_loading(self, model, **kwargs):
elif not self.is_quantization_compressed:
apply_quantization_config(model, ct_quantization_config)


def _process_model_after_weight_loading(self, model, **kwargs):
"""Decompress loaded model if necessary - need for qat"""
# enable post loading
Expand Down

0 comments on commit 1f6ed71

Please sign in to comment.