amogne-vlm-LLM / start.sh
amogneandualem's picture
Update start.sh
fa6a201 verified
raw
history blame contribute delete
827 Bytes
#!/bin/bash
# 1. Set the OLLAMA_HOST to listen on all interfaces (CRITICAL NETWORK FIX for Docker)
export OLLAMA_HOST=0.0.0.0
# 2. Start Ollama in the background
echo "Starting Ollama server in background (listening on $OLLAMA_HOST:11434)..."
ollama serve &
# 3. Wait for the Ollama server process to initialize
echo "Waiting 45 seconds for Ollama server to fully initialize..."
sleep 45
# 4. Pull and CONFIRM MODEL READINESS (Pre-load the model)
MODEL="moondream:1.8b"
echo "Ensuring model $MODEL is pulled and ready..."
# Pull and force a synchronous run to load the model into VRAM
ollama pull "$MODEL"
ollama run "$MODEL" "Describe a house" --verbose --system "You are a helpful assistant." --keep < /dev/null
# 5. Start the Gradio web application
echo "Starting Gradio application on port 7860..."
exec python app.py