add v1 model as default in custom monkey patch
This commit is contained in:
parent
f185b90c3e
commit
5d3267d80d
|
|
@ -16,7 +16,7 @@ def load_model_llama(*args, **kwargs):
|
|||
print("Loading {} ...".format(model_path))
|
||||
t0 = time.time()
|
||||
|
||||
model, tokenizer = load_llama_model_4bit_low_ram(config_path, model_path, groupsize=-1)
|
||||
model, tokenizer = load_llama_model_4bit_low_ram(config_path, model_path, groupsize=-1, is_v1_model=True)
|
||||
|
||||
model = PeftModel.from_pretrained(model, lora_path, device_map={'': 0}, torch_dtype=torch.float32)
|
||||
print('{} Lora Applied.'.format(lora_path))
|
||||
|
|
|
|||
Loading…
Reference in New Issue