Spaces:
Sleeping
Sleeping
Start adding AIA
Browse files- Dockerfile +33 -10
- nginx.conf +4 -0
- service-config.yml +3 -3
- start-services.sh +13 -3
Dockerfile
CHANGED
|
@@ -5,7 +5,13 @@ RUN useradd -m -u 1000 user
|
|
| 5 |
|
| 6 |
# Set environment variables
|
| 7 |
ENV HOME=/home/user \
|
| 8 |
-
PATH=/home/user/.local/bin:$PATH
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
|
| 10 |
# Install minimal dependencies
|
| 11 |
RUN apt-get update && apt-get install -y \
|
|
@@ -17,11 +23,22 @@ RUN apt-get update && apt-get install -y \
|
|
| 17 |
nginx \
|
| 18 |
&& rm -rf /var/lib/apt/lists/*
|
| 19 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 20 |
# Create directories and set permissions
|
| 21 |
RUN mkdir -p /tmp/llamafile && \
|
| 22 |
mkdir -p $HOME/models && \
|
| 23 |
mkdir -p $HOME/app && \
|
| 24 |
mkdir -p $HOME/app/docauth && \
|
|
|
|
| 25 |
mkdir -p $HOME/.llamafile && \
|
| 26 |
mkdir -p /var/cache/nginx && \
|
| 27 |
mkdir -p /var/log/nginx && \
|
|
@@ -43,17 +60,14 @@ RUN cd /tmp/llamafile && \
|
|
| 43 |
chmod +x /usr/local/bin/llamafiler && \
|
| 44 |
rm -rf /tmp/llamafile
|
| 45 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
# Copy nginx configuration
|
| 47 |
COPY nginx.conf /etc/nginx/nginx.conf
|
| 48 |
|
| 49 |
-
# Switch to non-root user
|
| 50 |
-
USER user
|
| 51 |
-
WORKDIR $HOME/app
|
| 52 |
-
|
| 53 |
-
# Download models
|
| 54 |
-
RUN wget -q https://huggingface.co/bartowski/gemma-2-2b-it-GGUF/resolve/main/gemma-2-2b-it-Q8_0.gguf -O $HOME/models/gemma-2b.gguf && \
|
| 55 |
-
wget -q https://huggingface.co/leliuga/all-MiniLM-L6-v2-GGUF/resolve/main/all-MiniLM-L6-v2.F16.gguf -O $HOME/models/embeddings.gguf
|
| 56 |
-
|
| 57 |
# Copy web files
|
| 58 |
COPY --chown=user:user index.html $HOME/app/docauth/
|
| 59 |
COPY --chown=user:user document-authoring.js $HOME/app/docauth/
|
|
@@ -63,6 +77,15 @@ COPY --chown=user:user Sample.docx $HOME/app/docauth/
|
|
| 63 |
COPY --chown=user:user start-services.sh $HOME/app/
|
| 64 |
RUN chmod +x $HOME/app/start-services.sh
|
| 65 |
|
| 66 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 67 |
|
| 68 |
CMD ["./start-services.sh"]
|
|
|
|
| 5 |
|
| 6 |
# Set environment variables
|
| 7 |
ENV HOME=/home/user \
|
| 8 |
+
PATH=/home/user/.local/bin:$PATH \
|
| 9 |
+
API_AUTH_TOKEN=secret \
|
| 10 |
+
JWT_ALGORITHM=RS256 \
|
| 11 |
+
DASHBOARD_USERNAME=dashboard \
|
| 12 |
+
DASHBOARD_PASSWORD=secret \
|
| 13 |
+
SECRET_KEY_BASE=secret-key-base \
|
| 14 |
+
JWT_PUBLIC_KEY="-----BEGIN PUBLIC KEY-----\nMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA2gzhmJ9TDanEzWdP1WG+\n0Ecwbe7f3bv6e5UUpvcT5q68IQJKP47AQdBAnSlFVi4X9SaurbWoXdS6jpmPpk24\nQvitzLNFphHdwjFBelTAOa6taZrSusoFvrtK9x5xsW4zzt/bkpUraNx82Z8MwLwr\nt6HlY7dgO9+xBAabj4t1d2t+0HS8O/ed3CB6T2lj6S8AbLDSEFc9ScO6Uc1XJlSo\nrgyJJSPCpNhSq3AubEZ1wMS1iEtgAzTPRDsQv50qWIbn634HLWxTP/UH6YNJBwzt\n3O6q29kTtjXlMGXCvin37PyX4Jy1IiPFwJm45aWJGKSfVGMDojTJbuUtM+8P9Rrn\nAwIDAQAB\n-----END PUBLIC KEY-----"
|
| 15 |
|
| 16 |
# Install minimal dependencies
|
| 17 |
RUN apt-get update && apt-get install -y \
|
|
|
|
| 23 |
nginx \
|
| 24 |
&& rm -rf /var/lib/apt/lists/*
|
| 25 |
|
| 26 |
+
# Install Node.js and pnpm
|
| 27 |
+
RUN curl -fsSL https://deb.nodesource.com/setup_20.x | bash - && \
|
| 28 |
+
apt-get update && \
|
| 29 |
+
apt-get install -y nodejs && \
|
| 30 |
+
npm install -g corepack && \
|
| 31 |
+
rm -rf /var/lib/apt/lists/*
|
| 32 |
+
|
| 33 |
+
# Enable pnpm
|
| 34 |
+
RUN corepack enable && corepack prepare pnpm@latest --activate
|
| 35 |
+
|
| 36 |
# Create directories and set permissions
|
| 37 |
RUN mkdir -p /tmp/llamafile && \
|
| 38 |
mkdir -p $HOME/models && \
|
| 39 |
mkdir -p $HOME/app && \
|
| 40 |
mkdir -p $HOME/app/docauth && \
|
| 41 |
+
mkdir -p $HOME/app/aia && \
|
| 42 |
mkdir -p $HOME/.llamafile && \
|
| 43 |
mkdir -p /var/cache/nginx && \
|
| 44 |
mkdir -p /var/log/nginx && \
|
|
|
|
| 60 |
chmod +x /usr/local/bin/llamafiler && \
|
| 61 |
rm -rf /tmp/llamafile
|
| 62 |
|
| 63 |
+
# Copy AI Assistant app
|
| 64 |
+
COPY --from=pspdfkit/ai-assistant:nightly /base $HOME/app/aia
|
| 65 |
+
COPY --chown=node:node service-config.yml /service-config.yml
|
| 66 |
+
RUN chown -R user:user $HOME/app/aia
|
| 67 |
+
|
| 68 |
# Copy nginx configuration
|
| 69 |
COPY nginx.conf /etc/nginx/nginx.conf
|
| 70 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 71 |
# Copy web files
|
| 72 |
COPY --chown=user:user index.html $HOME/app/docauth/
|
| 73 |
COPY --chown=user:user document-authoring.js $HOME/app/docauth/
|
|
|
|
| 77 |
COPY --chown=user:user start-services.sh $HOME/app/
|
| 78 |
RUN chmod +x $HOME/app/start-services.sh
|
| 79 |
|
| 80 |
+
# Switch to non-root user
|
| 81 |
+
USER user
|
| 82 |
+
WORKDIR $HOME/app
|
| 83 |
+
|
| 84 |
+
# Download models
|
| 85 |
+
RUN wget -q https://huggingface.co/bartowski/gemma-2-2b-it-GGUF/resolve/main/gemma-2-2b-it-Q8_0.gguf -O $HOME/models/gemma-2b.gguf && \
|
| 86 |
+
wget -q https://huggingface.co/leliuga/all-MiniLM-L6-v2-GGUF/resolve/main/all-MiniLM-L6-v2.F16.gguf -O $HOME/models/embeddings.gguf
|
| 87 |
+
|
| 88 |
+
# Expose (7860, for Hugging Face, 4000 for AI Assistant)
|
| 89 |
+
EXPOSE 7860
|
| 90 |
|
| 91 |
CMD ["./start-services.sh"]
|
nginx.conf
CHANGED
|
@@ -22,5 +22,9 @@ http {
|
|
| 22 |
location /v1/embeddings {
|
| 23 |
proxy_pass http://127.0.0.1:8081;
|
| 24 |
}
|
|
|
|
|
|
|
|
|
|
|
|
|
| 25 |
}
|
| 26 |
}
|
|
|
|
| 22 |
location /v1/embeddings {
|
| 23 |
proxy_pass http://127.0.0.1:8081;
|
| 24 |
}
|
| 25 |
+
|
| 26 |
+
location /v1 {
|
| 27 |
+
proxy_pass http://127.0.0.1:8082;
|
| 28 |
+
}
|
| 29 |
}
|
| 30 |
}
|
service-config.yml
CHANGED
|
@@ -4,17 +4,17 @@ aiServices:
|
|
| 4 |
chat:
|
| 5 |
provider:
|
| 6 |
name: 'openai-compat'
|
| 7 |
-
baseUrl: http://127.0.0.1:
|
| 8 |
model: 'gemma-2b'
|
| 9 |
textEmbeddings:
|
| 10 |
provider:
|
| 11 |
name: 'openai-compat'
|
| 12 |
-
baseUrl: http://127.0.0.1:
|
| 13 |
model: 'all-MiniLM-L6-v2'
|
| 14 |
headless:
|
| 15 |
- provider:
|
| 16 |
name: 'openai-compat'
|
| 17 |
-
baseUrl: http://127.0.0.1:
|
| 18 |
model:
|
| 19 |
name: 'gemma-2b'
|
| 20 |
id: 'gemma-2b'
|
|
|
|
| 4 |
chat:
|
| 5 |
provider:
|
| 6 |
name: 'openai-compat'
|
| 7 |
+
baseUrl: http://127.0.0.1:7861/v1
|
| 8 |
model: 'gemma-2b'
|
| 9 |
textEmbeddings:
|
| 10 |
provider:
|
| 11 |
name: 'openai-compat'
|
| 12 |
+
baseUrl: http://127.0.0.1:7861/v1
|
| 13 |
model: 'all-MiniLM-L6-v2'
|
| 14 |
headless:
|
| 15 |
- provider:
|
| 16 |
name: 'openai-compat'
|
| 17 |
+
baseUrl: http://127.0.0.1:7861/v1
|
| 18 |
model:
|
| 19 |
name: 'gemma-2b'
|
| 20 |
id: 'gemma-2b'
|
start-services.sh
CHANGED
|
@@ -25,6 +25,12 @@ if ! ps aux | grep nginx | grep -v grep > /dev/null; then
|
|
| 25 |
fi
|
| 26 |
echo "Nginx started successfully"
|
| 27 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 28 |
# Start the models
|
| 29 |
echo "Starting chat model..."
|
| 30 |
TMPDIR=/tmp/llamafiler /usr/local/bin/llamafiler --model $HOME/models/gemma-2b.gguf -ngl 999 --listen 0.0.0.0:8082 &
|
|
@@ -40,10 +46,10 @@ TIMEOUT=600 # 10 minutes timeout
|
|
| 40 |
START_TIME=$SECONDS
|
| 41 |
|
| 42 |
wait_for_services() {
|
| 43 |
-
curl -s --fail -X POST http://127.0.0.1:
|
| 44 |
-H "Content-Type: application/json" \
|
| 45 |
-d '{"model": "gemma-2b", "messages":[{"role":"user","content":"hi"}]}' >/dev/null 2>&1 && \
|
| 46 |
-
curl -s --fail -X POST http://127.0.0.1:
|
| 47 |
-H "Content-Type: application/json" \
|
| 48 |
-d '{"input":"test"}' >/dev/null 2>&1
|
| 49 |
}
|
|
@@ -64,8 +70,12 @@ until wait_for_services; do
|
|
| 64 |
echo "Embeddings model process died"
|
| 65 |
exit 1
|
| 66 |
fi
|
|
|
|
|
|
|
|
|
|
|
|
|
| 67 |
|
| 68 |
-
echo "Waiting for services... (${ELAPSED}s elapsed, PIDs: Gemma=$GEMMA_PID, Embeddings=$EMBEDDINGS_PID)"
|
| 69 |
sleep 2
|
| 70 |
done
|
| 71 |
|
|
|
|
| 25 |
fi
|
| 26 |
echo "Nginx started successfully"
|
| 27 |
|
| 28 |
+
# Start AI Assistant
|
| 29 |
+
echo "Starting AI Assistant..."
|
| 30 |
+
cd $HOME/app/aia
|
| 31 |
+
PORT=4000 node app/main.bundle.js &
|
| 32 |
+
AIA_PID=$!
|
| 33 |
+
|
| 34 |
# Start the models
|
| 35 |
echo "Starting chat model..."
|
| 36 |
TMPDIR=/tmp/llamafiler /usr/local/bin/llamafiler --model $HOME/models/gemma-2b.gguf -ngl 999 --listen 0.0.0.0:8082 &
|
|
|
|
| 46 |
START_TIME=$SECONDS
|
| 47 |
|
| 48 |
wait_for_services() {
|
| 49 |
+
curl -s --fail -X POST http://127.0.0.1:7861/v1/chat/completions \
|
| 50 |
-H "Content-Type: application/json" \
|
| 51 |
-d '{"model": "gemma-2b", "messages":[{"role":"user","content":"hi"}]}' >/dev/null 2>&1 && \
|
| 52 |
+
curl -s --fail -X POST http://127.0.0.1:7861/v1/embeddings \
|
| 53 |
-H "Content-Type: application/json" \
|
| 54 |
-d '{"input":"test"}' >/dev/null 2>&1
|
| 55 |
}
|
|
|
|
| 70 |
echo "Embeddings model process died"
|
| 71 |
exit 1
|
| 72 |
fi
|
| 73 |
+
if ! kill -0 $AIA_PID 2>/dev/null; then
|
| 74 |
+
echo "AI Assistant process died"
|
| 75 |
+
exit 1
|
| 76 |
+
fi
|
| 77 |
|
| 78 |
+
echo "Waiting for services... (${ELAPSED}s elapsed, PIDs: Gemma=$GEMMA_PID, Embeddings=$EMBEDDINGS_PID, AIA=$AIA_PID)"
|
| 79 |
sleep 2
|
| 80 |
done
|
| 81 |
|