#!/bin/bash # Exit on any error set -e # echo "Starting Ollama and FastAPI application..." # # Create necessary directories # mkdir -p /tmp/ollama/models # export OLLAMA_MODELS=/tmp/ollama/models # export OLLAMA_HOME=/tmp/ollama # # Start Ollama server in the background # echo "Starting Ollama server..." # ollama serve & # OLLAMA_PID=$! # Wait for Ollama to be ready # echo "Waiting for Ollama server to start..." # for i in {1..30}; do # if curl -s http://localhost:11434/api/tags > /dev/null 2>&1; then # echo "Ollama server is ready!" # break # fi # if [ $i -eq 30 ]; then # echo "Timeout waiting for Ollama server to start" # exit 1 # fi # sleep 2 # done # ollama list # Optional: Pull a model (uncomment and modify as needed) # echo "Pulling llama2 model..." # ollama pull tinyllama:1.1b # ollama pull HuggingFaceTB/SmolLM3-3B #deepseek-ai/DeepSeek-R1-0528-Qwen3-8B # Start FastAPI application echo "Starting FastAPI application..." python app.py # Keep the script running and handle cleanup cleanup() { echo "Shutting down..." # kill $OLLAMA_PID 2>/dev/null || true exit 0 } trap cleanup SIGTERM SIGINT # Wait for background processes wait