dmartincy commited on
Commit
a240151
·
1 Parent(s): b90ec9f

Start adding AIA

Browse files
Files changed (4) hide show
  1. Dockerfile +33 -10
  2. nginx.conf +4 -0
  3. service-config.yml +3 -3
  4. start-services.sh +13 -3
Dockerfile CHANGED
@@ -5,7 +5,13 @@ RUN useradd -m -u 1000 user
5
 
6
  # Set environment variables
7
  ENV HOME=/home/user \
8
- PATH=/home/user/.local/bin:$PATH
 
 
 
 
 
 
9
 
10
  # Install minimal dependencies
11
  RUN apt-get update && apt-get install -y \
@@ -17,11 +23,22 @@ RUN apt-get update && apt-get install -y \
17
  nginx \
18
  && rm -rf /var/lib/apt/lists/*
19
 
 
 
 
 
 
 
 
 
 
 
20
  # Create directories and set permissions
21
  RUN mkdir -p /tmp/llamafile && \
22
  mkdir -p $HOME/models && \
23
  mkdir -p $HOME/app && \
24
  mkdir -p $HOME/app/docauth && \
 
25
  mkdir -p $HOME/.llamafile && \
26
  mkdir -p /var/cache/nginx && \
27
  mkdir -p /var/log/nginx && \
@@ -43,17 +60,14 @@ RUN cd /tmp/llamafile && \
43
  chmod +x /usr/local/bin/llamafiler && \
44
  rm -rf /tmp/llamafile
45
 
 
 
 
 
 
46
  # Copy nginx configuration
47
  COPY nginx.conf /etc/nginx/nginx.conf
48
 
49
- # Switch to non-root user
50
- USER user
51
- WORKDIR $HOME/app
52
-
53
- # Download models
54
- RUN wget -q https://huggingface.co/bartowski/gemma-2-2b-it-GGUF/resolve/main/gemma-2-2b-it-Q8_0.gguf -O $HOME/models/gemma-2b.gguf && \
55
- wget -q https://huggingface.co/leliuga/all-MiniLM-L6-v2-GGUF/resolve/main/all-MiniLM-L6-v2.F16.gguf -O $HOME/models/embeddings.gguf
56
-
57
  # Copy web files
58
  COPY --chown=user:user index.html $HOME/app/docauth/
59
  COPY --chown=user:user document-authoring.js $HOME/app/docauth/
@@ -63,6 +77,15 @@ COPY --chown=user:user Sample.docx $HOME/app/docauth/
63
  COPY --chown=user:user start-services.sh $HOME/app/
64
  RUN chmod +x $HOME/app/start-services.sh
65
 
66
- EXPOSE 7860 8081
 
 
 
 
 
 
 
 
 
67
 
68
  CMD ["./start-services.sh"]
 
5
 
6
  # Set environment variables
7
  ENV HOME=/home/user \
8
+ PATH=/home/user/.local/bin:$PATH \
9
+ API_AUTH_TOKEN=secret \
10
+ JWT_ALGORITHM=RS256 \
11
+ DASHBOARD_USERNAME=dashboard \
12
+ DASHBOARD_PASSWORD=secret \
13
+ SECRET_KEY_BASE=secret-key-base \
14
+ JWT_PUBLIC_KEY="-----BEGIN PUBLIC KEY-----\nMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA2gzhmJ9TDanEzWdP1WG+\n0Ecwbe7f3bv6e5UUpvcT5q68IQJKP47AQdBAnSlFVi4X9SaurbWoXdS6jpmPpk24\nQvitzLNFphHdwjFBelTAOa6taZrSusoFvrtK9x5xsW4zzt/bkpUraNx82Z8MwLwr\nt6HlY7dgO9+xBAabj4t1d2t+0HS8O/ed3CB6T2lj6S8AbLDSEFc9ScO6Uc1XJlSo\nrgyJJSPCpNhSq3AubEZ1wMS1iEtgAzTPRDsQv50qWIbn634HLWxTP/UH6YNJBwzt\n3O6q29kTtjXlMGXCvin37PyX4Jy1IiPFwJm45aWJGKSfVGMDojTJbuUtM+8P9Rrn\nAwIDAQAB\n-----END PUBLIC KEY-----"
15
 
16
  # Install minimal dependencies
17
  RUN apt-get update && apt-get install -y \
 
23
  nginx \
24
  && rm -rf /var/lib/apt/lists/*
25
 
26
+ # Install Node.js and pnpm
27
+ RUN curl -fsSL https://deb.nodesource.com/setup_20.x | bash - && \
28
+ apt-get update && \
29
+ apt-get install -y nodejs && \
30
+ npm install -g corepack && \
31
+ rm -rf /var/lib/apt/lists/*
32
+
33
+ # Enable pnpm
34
+ RUN corepack enable && corepack prepare pnpm@latest --activate
35
+
36
  # Create directories and set permissions
37
  RUN mkdir -p /tmp/llamafile && \
38
  mkdir -p $HOME/models && \
39
  mkdir -p $HOME/app && \
40
  mkdir -p $HOME/app/docauth && \
41
+ mkdir -p $HOME/app/aia && \
42
  mkdir -p $HOME/.llamafile && \
43
  mkdir -p /var/cache/nginx && \
44
  mkdir -p /var/log/nginx && \
 
60
  chmod +x /usr/local/bin/llamafiler && \
61
  rm -rf /tmp/llamafile
62
 
63
+ # Copy AI Assistant app
64
+ COPY --from=pspdfkit/ai-assistant:nightly /base $HOME/app/aia
65
+ COPY --chown=node:node service-config.yml /service-config.yml
66
+ RUN chown -R user:user $HOME/app/aia
67
+
68
  # Copy nginx configuration
69
  COPY nginx.conf /etc/nginx/nginx.conf
70
 
 
 
 
 
 
 
 
 
71
  # Copy web files
72
  COPY --chown=user:user index.html $HOME/app/docauth/
73
  COPY --chown=user:user document-authoring.js $HOME/app/docauth/
 
77
  COPY --chown=user:user start-services.sh $HOME/app/
78
  RUN chmod +x $HOME/app/start-services.sh
79
 
80
+ # Switch to non-root user
81
+ USER user
82
+ WORKDIR $HOME/app
83
+
84
+ # Download models
85
+ RUN wget -q https://huggingface.co/bartowski/gemma-2-2b-it-GGUF/resolve/main/gemma-2-2b-it-Q8_0.gguf -O $HOME/models/gemma-2b.gguf && \
86
+ wget -q https://huggingface.co/leliuga/all-MiniLM-L6-v2-GGUF/resolve/main/all-MiniLM-L6-v2.F16.gguf -O $HOME/models/embeddings.gguf
87
+
88
+ # Expose (7860, for Hugging Face, 4000 for AI Assistant)
89
+ EXPOSE 7860
90
 
91
  CMD ["./start-services.sh"]
nginx.conf CHANGED
@@ -22,5 +22,9 @@ http {
22
  location /v1/embeddings {
23
  proxy_pass http://127.0.0.1:8081;
24
  }
 
 
 
 
25
  }
26
  }
 
22
  location /v1/embeddings {
23
  proxy_pass http://127.0.0.1:8081;
24
  }
25
+
26
+ location /v1 {
27
+ proxy_pass http://127.0.0.1:8082;
28
+ }
29
  }
30
  }
service-config.yml CHANGED
@@ -4,17 +4,17 @@ aiServices:
4
  chat:
5
  provider:
6
  name: 'openai-compat'
7
- baseUrl: http://127.0.0.1:8082/v1
8
  model: 'gemma-2b'
9
  textEmbeddings:
10
  provider:
11
  name: 'openai-compat'
12
- baseUrl: http://127.0.0.1:8081/v1
13
  model: 'all-MiniLM-L6-v2'
14
  headless:
15
  - provider:
16
  name: 'openai-compat'
17
- baseUrl: http://127.0.0.1:8082/v1
18
  model:
19
  name: 'gemma-2b'
20
  id: 'gemma-2b'
 
4
  chat:
5
  provider:
6
  name: 'openai-compat'
7
+ baseUrl: http://127.0.0.1:7861/v1
8
  model: 'gemma-2b'
9
  textEmbeddings:
10
  provider:
11
  name: 'openai-compat'
12
+ baseUrl: http://127.0.0.1:7861/v1
13
  model: 'all-MiniLM-L6-v2'
14
  headless:
15
  - provider:
16
  name: 'openai-compat'
17
+ baseUrl: http://127.0.0.1:7861/v1
18
  model:
19
  name: 'gemma-2b'
20
  id: 'gemma-2b'
start-services.sh CHANGED
@@ -25,6 +25,12 @@ if ! ps aux | grep nginx | grep -v grep > /dev/null; then
25
  fi
26
  echo "Nginx started successfully"
27
 
 
 
 
 
 
 
28
  # Start the models
29
  echo "Starting chat model..."
30
  TMPDIR=/tmp/llamafiler /usr/local/bin/llamafiler --model $HOME/models/gemma-2b.gguf -ngl 999 --listen 0.0.0.0:8082 &
@@ -40,10 +46,10 @@ TIMEOUT=600 # 10 minutes timeout
40
  START_TIME=$SECONDS
41
 
42
  wait_for_services() {
43
- curl -s --fail -X POST http://127.0.0.1:8082/v1/chat/completions \
44
  -H "Content-Type: application/json" \
45
  -d '{"model": "gemma-2b", "messages":[{"role":"user","content":"hi"}]}' >/dev/null 2>&1 && \
46
- curl -s --fail -X POST http://127.0.0.1:8081/v1/embeddings \
47
  -H "Content-Type: application/json" \
48
  -d '{"input":"test"}' >/dev/null 2>&1
49
  }
@@ -64,8 +70,12 @@ until wait_for_services; do
64
  echo "Embeddings model process died"
65
  exit 1
66
  fi
 
 
 
 
67
 
68
- echo "Waiting for services... (${ELAPSED}s elapsed, PIDs: Gemma=$GEMMA_PID, Embeddings=$EMBEDDINGS_PID)"
69
  sleep 2
70
  done
71
 
 
25
  fi
26
  echo "Nginx started successfully"
27
 
28
+ # Start AI Assistant
29
+ echo "Starting AI Assistant..."
30
+ cd $HOME/app/aia
31
+ PORT=4000 node app/main.bundle.js &
32
+ AIA_PID=$!
33
+
34
  # Start the models
35
  echo "Starting chat model..."
36
  TMPDIR=/tmp/llamafiler /usr/local/bin/llamafiler --model $HOME/models/gemma-2b.gguf -ngl 999 --listen 0.0.0.0:8082 &
 
46
  START_TIME=$SECONDS
47
 
48
  wait_for_services() {
49
+ curl -s --fail -X POST http://127.0.0.1:7861/v1/chat/completions \
50
  -H "Content-Type: application/json" \
51
  -d '{"model": "gemma-2b", "messages":[{"role":"user","content":"hi"}]}' >/dev/null 2>&1 && \
52
+ curl -s --fail -X POST http://127.0.0.1:7861/v1/embeddings \
53
  -H "Content-Type: application/json" \
54
  -d '{"input":"test"}' >/dev/null 2>&1
55
  }
 
70
  echo "Embeddings model process died"
71
  exit 1
72
  fi
73
+ if ! kill -0 $AIA_PID 2>/dev/null; then
74
+ echo "AI Assistant process died"
75
+ exit 1
76
+ fi
77
 
78
+ echo "Waiting for services... (${ELAPSED}s elapsed, PIDs: Gemma=$GEMMA_PID, Embeddings=$EMBEDDINGS_PID, AIA=$AIA_PID)"
79
  sleep 2
80
  done
81