update multi gpu support in finetune.py
This commit is contained in:
parent
5655f218ed
commit
86387a0a35
|
|
@ -59,7 +59,10 @@ lora_config = LoraConfig(
|
||||||
if ft_config.lora_apply_dir is None:
|
if ft_config.lora_apply_dir is None:
|
||||||
model = get_peft_model(model, lora_config)
|
model = get_peft_model(model, lora_config)
|
||||||
else:
|
else:
|
||||||
model = PeftModel.from_pretrained(model, ft_config.lora_apply_dir, device_map={'': 0}, torch_dtype=torch.float32) # ! Direct copy from inference.py
|
if ft_config.ddp:
|
||||||
|
model = PeftModel.from_pretrained(model, ft_config.lora_apply_dir, device_map="auto", torch_dtype=torch.float32) # ! Direct copy from inference.py
|
||||||
|
else:
|
||||||
|
model = PeftModel.from_pretrained(model, ft_config.lora_apply_dir, device_map={'': 0}, torch_dtype=torch.float32)
|
||||||
print(ft_config.lora_apply_dir, 'loaded')
|
print(ft_config.lora_apply_dir, 'loaded')
|
||||||
|
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue