From 13028b8d2b82abf9778db97d8c378647aec5defb Mon Sep 17 00:00:00 2001 From: crowetic Date: Mon, 19 May 2025 14:05:34 -0700 Subject: [PATCH] tweaked localai config a bit more --- setup-ai-stack.sh | 12 ++++++++++-- 1 file changed, 10 insertions(+), 2 deletions(-) diff --git a/setup-ai-stack.sh b/setup-ai-stack.sh index 98937dc..b177b8f 100644 --- a/setup-ai-stack.sh +++ b/setup-ai-stack.sh @@ -73,7 +73,11 @@ docker run -d \ --gpus all \ -p 8080:8080 \ -v ~/ai-stack/localai/models:/models \ - --env ENABLE_BACKENDS=llama-cuda,ollama \ + -v ~/ai-stack/localai/config:/config \ + -e ENABLE_BACKENDS=llama-cuda,ollama \ + -e INCLUDE_DEFAULT_MODELS=true \ + -e AUTOLOAD_MODELS=true \ + -e MODEL_PATH=/models \ --restart unless-stopped \ localai/localai:latest-aio-gpu-nvidia-cuda-12 @@ -439,7 +443,11 @@ EOF --gpus all \ -p 8080:8080 \ -v ~/ai-stack/localai/models:/models \ - --env ENABLE_BACKENDS=llama-cuda,ollama \ + -v ~/ai-stack/localai/config:/config \ + -e ENABLE_BACKENDS=llama-cuda,ollama \ + -e INCLUDE_DEFAULT_MODELS=true \ + -e AUTOLOAD_MODELS=true \ + -e MODEL_PATH=/models \ --restart unless-stopped \ localai/localai:latest-aio-gpu-nvidia-cuda-12