--- license: apache-2.0 tags: - llama - gguf - quantized library_name: transformers --- # ⚠️ I apologize for not providing any files here. This is just a generated text. # TinyLlama PHP Fine-tuned GGUF This is a GGUF conversion of the TinyLlama model fine-tuned for PHP code generation. ## Model Details - **Base Model**: TinyLlama - **Fine-tuned for**: PHP code generation - **Format**: GGUF (quantized to q4_0) - **Use with**: llama.cpp, Ollama, or other GGUF-compatible inference engines ## Usage ### With llama.cpp: ```bash ./main -m model.gguf -p "Write a PHP function to" ``` ### With Ollama: 1. Create a Modelfile: ``` FROM ./model.gguf ``` 2. Create the model: ```bash ollama create tinyllama-php -f Modelfile ollama run tinyllama-php ```