Files
comfyui-nvidia/deployments/ai-stack/init-models.sh
William Gill 5a34ced8f1 Add S3 mirror path for Ollama models + mirror-ollama-model.sh helper
Three pieces:

1. mirror-ollama-model.sh — run on any machine that has the model
   pulled. Parses the manifest at
   ~/.ollama/models/manifests/registry.ollama.ai/<ns>/<name>/<tag>,
   greps every sha256:* digest, tars manifest + referenced blobs into
   one .tgz. Output is portable — extract over any other Ollama
   data dir and the model is immediately visible.

2. init-models.sh gains an s3_pull function that curls a tarball from
   $S3_OLLAMA_BASE and extracts into /root/.ollama/models/. Falls back
   to ollama pull when S3_OLLAMA_BASE is unset, so s3_pull lines are
   safe to commit before the bucket is ready. huihui_ai/qwen3.5-
   abliterated:9b promoted to s3_pull as the example.

3. docker-compose.yml model-init service propagates S3_OLLAMA_BASE
   from .env. Curl auto-installs at script start because ollama/ollama
   doesn't always ship it.

README documents the mirror workflow under "Mirroring models to S3".

Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
2026-04-19 13:43:26 -05:00

64 lines
2.4 KiB
Bash

#!/bin/sh
# Preseed Ollama with the models the stack should have available at startup.
# Runs once via the model-init service (see docker-compose.yml). Safe to
# re-run — already-present models are skipped.
#
# Two pull paths:
# - s3_pull — fetches a tarball from $S3_OLLAMA_BASE (your own mirror,
# created by mirror-ollama-model.sh) and extracts into
# Ollama's data dir. Faster + immune to upstream changes.
# Falls back to ollama pull if S3_OLLAMA_BASE is unset.
# - pull — standard `ollama pull` against registry.ollama.ai.
set -e
# Make sure curl is available — ollama/ollama:latest doesn't always include
# it, and s3_pull needs it. tar is in the base image.
if ! command -v curl >/dev/null 2>&1; then
apt-get update -qq && apt-get install -y -qq curl ca-certificates >/dev/null
fi
S3_OLLAMA_BASE="${S3_OLLAMA_BASE:-}"
OLLAMA_DATA="/root/.ollama"
s3_pull() {
name="$1"; archive="$2"
if ollama list 2>/dev/null | awk 'NR>1 {print $1}' | grep -qx "$name"; then
echo "$name already present"
return
fi
if [ -z "$S3_OLLAMA_BASE" ]; then
echo "$name: S3_OLLAMA_BASE unset, falling back to ollama pull"
ollama pull "$name"
return
fi
url="${S3_OLLAMA_BASE%/}/$archive"
echo "→ Downloading $name from $url"
curl -fL -C - --retry 3 -o "/tmp/$archive" "$url"
tar -xzf "/tmp/$archive" -C "$OLLAMA_DATA/models/"
rm -f "/tmp/$archive"
echo "$name installed (mirror)"
}
pull() {
name="$1"
if ollama list 2>/dev/null | awk 'NR>1 {print $1}' | grep -qx "$name"; then
echo "$name already present"
else
echo "→ Pulling $name from registry.ollama.ai…"
ollama pull "$name"
fi
}
# ─── S3-mirrored models ─────────────────────────────────────────────────────
# These live in your own bucket. Create the tarballs once with
# mirror-ollama-model.sh, upload to S3, then list them here.
s3_pull "huihui_ai/qwen3.5-abliterated:9b" "qwen3.5-abliterated-9b.tgz"
# ─── Direct registry pulls ──────────────────────────────────────────────────
for model in dolphin3:8b llama3.1:8b ministral-3:8b mistral-nemo:12b qwen3.6:latest; do
pull "$model"
done
echo "Done."