UP
This commit is contained in:
parent
6c9c3f663b
commit
eb06e8bdce
17
Dockerfile
17
Dockerfile
@ -1,5 +1,8 @@
|
|||||||
FROM quay.io/go-skynet/local-ai:latest
|
FROM quay.io/go-skynet/local-ai:latest
|
||||||
|
|
||||||
|
# Utiliser bash pour les heredocs propres
|
||||||
|
SHELL ["/bin/bash", "-lc"]
|
||||||
|
|
||||||
ENV MODELS_PATH=/models \
|
ENV MODELS_PATH=/models \
|
||||||
DEBUG=false \
|
DEBUG=false \
|
||||||
THREADS=2 \
|
THREADS=2 \
|
||||||
@ -15,16 +18,16 @@ RUN curl -L --fail --progress-bar \
|
|||||||
-o /models/tinyllama-1.1b-chat.Q4_K_M.gguf \
|
-o /models/tinyllama-1.1b-chat.Q4_K_M.gguf \
|
||||||
https://huggingface.co/TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF/resolve/main/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf
|
https://huggingface.co/TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF/resolve/main/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf
|
||||||
|
|
||||||
# Mappe le nom du modèle **gpt-oss-20b** (compat) vers le fichier téléchargé
|
# Génère le mapping YAML du modèle exposé sous le nom "gpt-oss-20b"
|
||||||
# => côté Symfony on référence simplement "gpt-oss-20b"
|
RUN cat > /models/gpt-oss-20b.yaml <<'YAML'
|
||||||
RUN printf "name: gpt-oss-20b
|
name: gpt-oss-20b
|
||||||
backend: llama
|
backend: llama
|
||||||
parameters:
|
parameters:
|
||||||
model: tinyllama-1.1b-chat.Q4_K_M.gguf
|
model: tinyllama-1.1b-chat.Q4_K_M.gguf
|
||||||
n_ctx: %s
|
n_ctx: ${CONTEXT_SIZE}
|
||||||
n_threads: %s
|
n_threads: ${THREADS}
|
||||||
temperature: 0.2
|
temperature: 0.2
|
||||||
top_p: 0.9
|
top_p: 0.9
|
||||||
" "$CONTEXT_SIZE" "$THREADS" > /models/gpt-oss-20b.yaml
|
YAML
|
||||||
|
|
||||||
# L'entrée du serveur (déjà gérée par docker-compose via `command`)
|
# Le démarrage du serveur est piloté par docker-compose via `command`
|
||||||
Loading…
Reference in New Issue
Block a user