From 1f6ed71f6fa246cd75a086206b68ac47f1e66f8d Mon Sep 17 00:00:00 2001 From: Dipika Date: Thu, 16 Jan 2025 23:08:38 +0000 Subject: [PATCH] style --- src/transformers/quantizers/quantizer_compressed_tensors.py | 2 -- 1 file changed, 2 deletions(-) diff --git a/src/transformers/quantizers/quantizer_compressed_tensors.py b/src/transformers/quantizers/quantizer_compressed_tensors.py index afff4f59d73fee..9565bd366557c1 100644 --- a/src/transformers/quantizers/quantizer_compressed_tensors.py +++ b/src/transformers/quantizers/quantizer_compressed_tensors.py @@ -55,7 +55,6 @@ def __init__(self, quantization_config: CompressedTensorsConfig, **kwargs): self.weight_norm = torch.nn.utils.parametrizations.weight_norm delattr(torch.nn.utils.parametrizations, "weight_norm") - def validate_environment(self, *args, **kwargs): if not is_compressed_tensors_available(): raise ImportError( @@ -86,7 +85,6 @@ def _process_model_before_weight_loading(self, model, **kwargs): elif not self.is_quantization_compressed: apply_quantization_config(model, ct_quantization_config) - def _process_model_after_weight_loading(self, model, **kwargs): """Decompress loaded model if necessary - need for qat""" # enable post loading