|
FROM ubuntu:22.04
|
|
|
|
|
|
RUN apt-get update && \
|
|
apt-get install -y \
|
|
build-essential \
|
|
libssl-dev \
|
|
zlib1g-dev \
|
|
libboost-system-dev \
|
|
libboost-filesystem-dev \
|
|
cmake \
|
|
git \
|
|
python3-pip \
|
|
curl \
|
|
wget && \
|
|
rm -rf /var/lib/apt/lists/*
|
|
|
|
|
|
RUN pip3 install huggingface-hub openai gradio
|
|
|
|
|
|
RUN git clone https://github.com/ggerganov/llama.cpp && \
|
|
cd llama.cpp && \
|
|
mkdir build && \
|
|
cd build && \
|
|
cmake .. -DLLAMA_BUILD_SERVER=ON -DLLAMA_BUILD_EXAMPLES=ON -DCMAKE_BUILD_TYPE=Release && \
|
|
cmake --build . --config Release --target llama-server -j $(nproc)
|
|
|
|
|
|
RUN mkdir -p /models && \
|
|
wget -O /models/model.q2_k_l.gguf https://huggingface.co/unsloth/DeepSeek-R1-Distill-Qwen-32B-GGUF/resolve/main/DeepSeek-R1-Distill-Qwen-32B-Q2_K_L.gguf
|
|
|
|
|
|
COPY app.py /app.py
|
|
COPY start.sh /start.sh
|
|
RUN chmod +x /start.sh
|
|
|
|
|
|
EXPOSE 7860 8080
|
|
|
|
|
|
CMD ["/start.sh"] |