# if from main: | |
# (base) ubuntu@compute-permanent-node-406:~/lmdeploy$ docker build . -f docker/Dockerfile -t internvlmain --no-cache | |
# then change below "FROM openmmlab/lmdeploy:latest" to "FROM internvlmain" | |
# docker build - < Dockerfile.internvl -t internvl | |
FROM openmmlab/lmdeploy:latest | |
RUN apt-get update && apt-get install -y python3 python3-pip git | |
WORKDIR /app | |
RUN pip3 uninstall pkg_resources -y | |
RUN pip3 install --upgrade pip | |
RUN pip3 install --upgrade setuptools==66.1.1 | |
RUN pip3 uninstall -y ninja && pip3 install ninja | |
RUN CUDA_HOME=/usr/local/cuda-11.8/ PIP_EXTRA_INDEX_URL=https://download.pytorch.org/whl/cu118 pip3 install timm xformers triton==2.1.0 transformers | |
RUN MAX_JOBS=4 CUDA_HOME=/usr/local/cuda-11.8/ PIP_EXTRA_INDEX_URL=https://download.pytorch.org/whl/cu118 FLASH_ATTENTION_FORCE_BUILD=TRUE pip3 install flash-attn==2.5.2 --no-build-isolation | |
RUN pip3 install git+https://github.com/haotian-liu/LLaVA.git --no-deps | |
COPY . . | |
CMD ["lmdeploy", "serve", "api_server", "OpenGVLab/InternVL-Chat-V1-5"] | |