alpaca_lora_4bit/GPTQ-for-LLaMa
John Smith 04f5575a23
reduced memory usage by a little
2023-03-20 00:51:52 +08:00
..
autograd_4bit.py reduced memory usage by a little 2023-03-20 00:51:52 +08:00
quant_cuda.cpp add patch for gptq and peft 2023-03-18 13:31:48 +08:00
quant_cuda_kernel.cu add patch for gptq and peft 2023-03-18 13:31:48 +08:00