From 00bf0a1e1b205fa706bf4020490cba2d568ed9c2 Mon Sep 17 00:00:00 2001 From: John Smith Date: Fri, 31 Mar 2023 14:17:35 +0800 Subject: [PATCH] Update README.md --- README.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/README.md b/README.md index 21bfc9b..2326cde 100644 --- a/README.md +++ b/README.md @@ -16,9 +16,9 @@ Made some adjust for the code in peft and gptq for llama, and make it possible f * Added offload support. load_llama_model_4bit_low_ram_and_offload_to_cpu function can be used. # Requirements -gptq-for-llama: https://github.com/qwopqwop200/GPTQ-for-LLaMa
-peft: https://github.com/huggingface/peft.git
-
+gptq-for-llama
+peft
+The specific version is inside requirements.txt
# Install ~copy files from GPTQ-for-LLaMa into GPTQ-for-LLaMa path and re-compile cuda extension~