|
FROM python:3.12 |
|
|
|
RUN useradd -m -u 1000 app |
|
|
|
WORKDIR /home/app |
|
|
|
RUN apt update |
|
RUN apt install -y wget make cmake clang git g++ |
|
RUN wget https://huggingface.co/neph1/llama-3.1-instruct-bellman-8b-swedish/resolve/main/gguf/llama-3.1-instruct-bellman-8b-swedish-Q8.gguf?download=true -O model.gguf |
|
RUN git clone https://github.com/ggerganov/llama.cpp |
|
RUN mv llama.cpp llama_temp |
|
RUN mv llama_temp/* . |
|
RUN make |
|
RUN apt install socat -y |
|
|
|
EXPOSE 7860 |
|
|
|
CMD ["sh", "-c", "./llama-server -m /home/app/model.gguf -c 8192 --host 0.0.0.0 --port 7860"] |