From 4c24344796db1a17116964027da9bb71c503735d Mon Sep 17 00:00:00 2001 From: Atinoda <61033436+Atinoda@users.noreply.github.com> Date: Tue, 27 Jun 2023 19:44:08 +0100 Subject: [PATCH] Remove ExLlama manual installation (no longer required) --- Dockerfile | 3 --- 1 file changed, 3 deletions(-) diff --git a/Dockerfile b/Dockerfile index caade7b..7df0f50 100644 --- a/Dockerfile +++ b/Dockerfile @@ -34,9 +34,6 @@ RUN git clone https://github.com/oobabooga/GPTQ-for-LLaMa.git -b cuda /app/repos # Build and install default GPTQ ('quant_cuda') ARG TORCH_CUDA_ARCH_LIST="6.1;7.0;7.5;8.0;8.6+PTX" RUN cd /app/repositories/GPTQ-for-LLaMa/ && python3 setup_cuda.py install -# Install ExLlama -RUN pip install safetensors sentencepiece ninja && \ - cd /app/repositories && git clone https://github.com/turboderp/exllama FROM nvidia/cuda:11.8.0-devel-ubuntu22.04 AS base # Runtime pre-reqs