--- base_model: - google/gemma-3-270m --- # gemma-3-270m GGUF Recommended way to run this model: ```sh llama-cli -hf ggml-org/gemma-3-270m-GGUF -c 0 -fa -p "hello" ``` Then, access http://localhost:8080