New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Docker not using GPU #746
Comments
I have the same problem. Any ideas? |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Hello,
I know this topic may have been mentioned before, but unfortunately, nothing has worked for me.
On Windows, I've never been able to get the models to work with my GPU (except when using text gen webui for another project). So, I switched to Docker, but unfortunately, I still have the same problem: BLAS = 0!
I have an Nvidia 3070 with 8GB of VRAM. Here's my Dockerfile:
`
syntax=docker/dockerfile:1
Build as
docker build . -t localgpt
, requires BuildKit.Run as
docker run -it --mount src="$HOME/.cache",target=/root/.cache,type=bind --gpus=all localgpt
, requires Nvidia container toolkit.FROM nvidia/cuda:11.7.1-runtime-ubuntu22.04
RUN apt-get update && apt-get install -y software-properties-common && apt-get install ffmpeg libsm6 libxext6 -y
RUN apt-get install -y g++-11 make python3 python-is-python3 pip
only copy what's needed at every step to optimize layer cache
COPY ./requirements.txt .
use BuildKit cache mount to drastically reduce redownloading from pip on repeated builds
RUN --mount=type=cache,target=/root/.cache CMAKE_ARGS="-DLLAMA_CUBLAS=on" FORCE_CMAKE=1 pip install --timeout 100 -r requirements.txt llama-cpp-python==0.1.83
COPY SOURCE_DOCUMENTS ./SOURCE_DOCUMENTS
COPY ingest.py constants.py ./
Docker BuildKit does not support GPU during docker build time right now, only during docker run.
See moby/buildkit#1436.
If this changes in the future you can
docker build --build-arg device_type=cuda . -t localgpt
(+GPU argument to be determined).ARG device_type=cpu
RUN --mount=type=cache,target=/root/.cache python ingest.py --device_type $device_type
COPY . .
ENV device_type=cuda
#CMD python run_localGPT.py --device_type $device_type
`
I'm trying to get Mistral gguf to work. I know there's a way to make it work on both GPU and CPU (at least with text gen webui, that's the case).
Thank you in advance, and I hope someone will have the time to help me :)
The text was updated successfully, but these errors were encountered: