From 6f4bbb40a9071baf17c05ed711b288eb06546590 Mon Sep 17 00:00:00 2001 From: John Smith Date: Sat, 18 Mar 2023 18:49:26 +0800 Subject: [PATCH] Update autograd_4bit.py --- GPTQ-for-LLaMa/autograd_4bit.py | 5 ----- 1 file changed, 5 deletions(-) diff --git a/GPTQ-for-LLaMa/autograd_4bit.py b/GPTQ-for-LLaMa/autograd_4bit.py index 0837a1d..a56bb1c 100644 --- a/GPTQ-for-LLaMa/autograd_4bit.py +++ b/GPTQ-for-LLaMa/autograd_4bit.py @@ -115,11 +115,6 @@ def load_llama_model_4bit_low_ram(config_path, model_path): with accelerate.init_empty_weights(): config = LLaMAConfig.from_pretrained(config_path) - def noop(*args, **kwargs): - pass - torch.nn.init.kaiming_uniform_ = noop - torch.nn.init.uniform_ = noop - torch.nn.init.normal_ = noop torch.set_default_dtype(torch.half) transformers.modeling_utils._init_weights = False torch.set_default_dtype(torch.half)