#!/bin/bash if [ ! -d "llama.cpp" ]; then # only run in dev env git clone https://github.com/ggerganov/llama.cpp fi if [[ -z "${RUN_LOCALLY}" ]]; then # enable CUDA if NOT running locally export LLAMA_CUDA=1 fi cd llama.cpp make -j llama-quantize llama-gguf-split llama-imatrix cd .. python app.py