From 3de5ecd7333779cb6d0e260c1c37a905470b5734 Mon Sep 17 00:00:00 2001 From: Luke Marsden Date: Wed, 16 Oct 2024 11:34:30 +0100 Subject: [PATCH] simplify the installer, we no longer need to specify RUNTIME_OLLAMA_WARMUP_MODELS because the code defaults to the models in -small, and -large will work without specifying them explicitly because they're already downloaded --- install.sh | 13 +------------ 1 file changed, 1 insertion(+), 12 deletions(-) diff --git a/install.sh b/install.sh index c0dff39f..3d38a940 100755 --- a/install.sh +++ b/install.sh @@ -711,13 +711,11 @@ if [ "$RUNNER" = true ]; then read -p "Please specify the GPU memory in GB: " GPU_MEMORY fi - # Determine runner tag and warmup models + # Determine runner tag if [ "$LARGE" = true ]; then RUNNER_TAG="${LATEST_RELEASE}-large" - WARMUP_MODELS="" else RUNNER_TAG="${LATEST_RELEASE}-small" - WARMUP_MODELS="llama3:instruct,phi3:instruct" fi # Determine runner token @@ -739,18 +737,10 @@ if [ "$RUNNER" = true ]; then RUNNER_TAG="${RUNNER_TAG}" API_HOST="${API_HOST}" GPU_MEMORY="${GPU_MEMORY}" -WARMUP_MODELS="${WARMUP_MODELS}" RUNNER_TOKEN="${RUNNER_TOKEN}" OLDER_GPU="${OLDER_GPU:-false}" HF_TOKEN="${HF_TOKEN}" -# Set warmup models parameter -if [ -n "\$WARMUP_MODELS" ]; then - WARMUP_MODELS_PARAM="-e RUNTIME_OLLAMA_WARMUP_MODELS=\$WARMUP_MODELS" -else - WARMUP_MODELS_PARAM="" -fi - # Set older GPU parameter if [ "\$OLDER_GPU" = "true" ]; then OLDER_GPU_PARAM="-e RUNTIME_AXOLOTL_ENABLED=false" @@ -786,7 +776,6 @@ sudo docker run --privileged --gpus all --shm-size=10g \\ --ulimit stack=67108864 \\ --network="helix_default" \\ -v \${HOME}/.cache/huggingface:/root/.cache/huggingface \\ - \${WARMUP_MODELS_PARAM} \\ \${OLDER_GPU_PARAM} \\ \${HF_TOKEN_PARAM} \\ registry.helix.ml/helix/runner:\${RUNNER_TAG} \\