unclemusclez commited on
Commit
fa322c3
·
verified ·
1 Parent(s): 7d8c66b

Update Dockerfile

Browse files
Files changed (1) hide show
  1. Dockerfile +20 -8
Dockerfile CHANGED
@@ -22,6 +22,7 @@ RUN apt-get update && \
22
  libxmlsec1-dev \
23
  libffi-dev \
24
  liblzma-dev \
 
25
  golang-1.22-go \
26
  nvidia-driver-550 \
27
  ffmpeg
@@ -30,9 +31,10 @@ ENV USER='user'
30
  RUN useradd -m -u 1000 ${USER}
31
  USER ${USER}
32
  ENV HOME=/home/${USER} \
33
- PATH=${HOME}/.local/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:${PATH}
34
- WORKDIR ${HOME}/app
35
-
 
36
 
37
  ENV NVIDIA_VISIBLE_DEVICES=all
38
  RUN curl https://pyenv.run | bash
@@ -44,14 +46,12 @@ RUN pyenv install ${PYTHON_VERSION} && \
44
  pip install --no-cache-dir -U pip setuptools wheel && \
45
  pip install "huggingface-hub" "hf-transfer" "gradio[oauth]>=4.28.0" "gradio_huggingfacehub_search==0.0.7" "APScheduler"
46
 
47
- COPY --chown=1000 . ${HOME}/app
48
-
49
  RUN git clone https://github.com/ollama/ollama
50
  RUN git clone https://github.com/ggerganov/llama.cpp
51
- RUN pip install -r llama.cpp/requirements.txt
52
  COPY groups_merged.txt ${HOME}/app/llama.cpp/.
53
 
54
- ENV PYTHONPATH=${HOME}/app \
55
  PYTHONUNBUFFERED=1 \
56
  HF_HUB_ENABLE_HF_TRANSFER=1 \
57
  GRADIO_ALLOW_FLAGGING=never \
@@ -66,5 +66,17 @@ ENV PYTHONPATH=${HOME}/app \
66
  NVIDIA_VISIBLE_DEVICES=all \
67
  OLLAMA_HOST=0.0.0.0
68
 
 
 
 
 
 
 
 
 
 
 
 
69
  # EXPOSE map[11434/tcp:{}]
70
- ENTRYPOINT /bin/sh start.sh
 
 
22
  libxmlsec1-dev \
23
  libffi-dev \
24
  liblzma-dev \
25
+ golang-go \
26
  golang-1.22-go \
27
  nvidia-driver-550 \
28
  ffmpeg
 
31
  RUN useradd -m -u 1000 ${USER}
32
  USER ${USER}
33
  ENV HOME=/home/${USER} \
34
+ PATH=${HOME}/.local/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:${PATH} \
35
+ APPDIR=${HOME}/app
36
+
37
+ WORKDIR ${APPDIR}
38
 
39
  ENV NVIDIA_VISIBLE_DEVICES=all
40
  RUN curl https://pyenv.run | bash
 
46
  pip install --no-cache-dir -U pip setuptools wheel && \
47
  pip install "huggingface-hub" "hf-transfer" "gradio[oauth]>=4.28.0" "gradio_huggingfacehub_search==0.0.7" "APScheduler"
48
 
49
+ COPY --chown=1000 . ${APPDIR}
 
50
  RUN git clone https://github.com/ollama/ollama
51
  RUN git clone https://github.com/ggerganov/llama.cpp
 
52
  COPY groups_merged.txt ${HOME}/app/llama.cpp/.
53
 
54
+ ENV PYTHONPATH=${APPDIR} \
55
  PYTHONUNBUFFERED=1 \
56
  HF_HUB_ENABLE_HF_TRANSFER=1 \
57
  GRADIO_ALLOW_FLAGGING=never \
 
66
  NVIDIA_VISIBLE_DEVICES=all \
67
  OLLAMA_HOST=0.0.0.0
68
 
69
+ WORKDIR ${APPDIR}/ollama
70
+ RUN OLLAMA_CUSTOM_CPU_DEFS="-DGGML_AVX=on -DGGML_AVX2=on -DGGML_F16C=on -DGGML_FMA=on" go generate ./... --verbose \
71
+ go build . --verbose \
72
+ go install . --verbose
73
+
74
+ WORKDIR ${APPDIR}/llama.cpp
75
+ RUN pip install -r requirements.txt \
76
+ LLAMA_CUDA=0 make -j llama-quantize --verbose
77
+
78
+ WORKDIR ${APPDIR}
79
+
80
  # EXPOSE map[11434/tcp:{}]
81
+ ENTRYPOINT ollama serve --verbose & sleep 5 && \
82
+ python app.py --verbose