Fix some issues.

This commit is contained in:
Andy Barry 2023-04-05 23:29:10 -04:00
parent 417eba372a
commit 1af4023883
2 changed files with 6 additions and 12 deletions

View File

@ -12,8 +12,7 @@ RUN pip3 install --upgrade pip requests tqdm
RUN pip install torch==2.0.0 RUN pip install torch==2.0.0
RUN pip install semantic-version==2.10.0 RUN pip install semantic-version==2.10.0
RUN git clone --depth=1 --branch main https://github.com/andybarry/alpaca_lora_4bit_docker.git && cd alpaca_lora_4bit RUN git clone --depth=1 --branch main https://github.com/andybarry/alpaca_lora_4bit_docker.git alpaca_lora_4bit && cd alpaca_lora_4bit
# && git checkout 86387a0a3575c82e689a452c20b2c9a5cc94a0f3
WORKDIR alpaca_lora_4bit WORKDIR alpaca_lora_4bit
@ -26,7 +25,6 @@ COPY requirements.txt requirements.txt
RUN pip install -r requirements.txt RUN pip install -r requirements.txt
RUN git clone --depth=1 --branch main https://github.com/andybarry/text-generation-webui-4bit.git text-generation-webui-tmp && cd text-generation-webui-tmp RUN git clone --depth=1 --branch main https://github.com/andybarry/text-generation-webui-4bit.git text-generation-webui-tmp && cd text-generation-webui-tmp
# && git checkout 378d21e80c3d6f11a4835e57597c69e340008e2c
RUN mv -f text-generation-webui-tmp/* text-generation-webui/ RUN mv -f text-generation-webui-tmp/* text-generation-webui/
@ -43,4 +41,4 @@ RUN cd text-generation-webui && ln -s ../autograd_4bit.py ./autograd_4bit.py &&
# Run the server # Run the server
WORKDIR /alpaca_lora_4bit/text-generation-webui WORKDIR /alpaca_lora_4bit/text-generation-webui
CMD ["python", "server.py"] CMD ["python", "-u", "server.py", "--listen", "--chat"]

View File

@ -7,8 +7,8 @@ Based on https://github.com/johnsmith0031/alpaca_lora_4bit
Can run real-time LLM chat using alpaca on a 8GB NVIDIA/CUDA GPU (ie 3070 Ti mobile) Can run real-time LLM chat using alpaca on a 8GB NVIDIA/CUDA GPU (ie 3070 Ti mobile)
## Requirements ## Requirements
- linux with docker - Docker
- nvidia GPU - NVIDIA GPU
## Installation ## Installation
@ -19,13 +19,9 @@ docker run -p 7086:7086 alpaca_lora_4bit
Point your browser to http://localhost:7086 Point your browser to http://localhost:7086
## Results ## Results
It's fast on a 3070 Ti. It's fast on a 3070 Ti mobile. Uses 5-6 GB of GPU RAM.
### Discussion
The model isn't all that good, sometimes it goes crazy. But hey, "when 4-bits _you reach_ look this good you will not."
But it is fast (on my 3070 Ti mobile at least)
The model isn't all that good, sometimes it goes crazy. But hey, as I always say, "when 4-bits _you reach_ look this good you will not."
## References ## References