nufikq's picture
Update README.md
31f4adc verified
metadata
license: apache-2.0
tags:
  - llama
  - gguf
  - quantized
library_name: transformers

⚠️ I apologize for not providing any files here. This is just a generated text.

TinyLlama PHP Fine-tuned GGUF

This is a GGUF conversion of the TinyLlama model fine-tuned for PHP code generation.

Model Details

  • Base Model: TinyLlama
  • Fine-tuned for: PHP code generation
  • Format: GGUF (quantized to q4_0)
  • Use with: llama.cpp, Ollama, or other GGUF-compatible inference engines

Usage

With llama.cpp:

./main -m model.gguf -p "Write a PHP function to"

With Ollama:

  1. Create a Modelfile:
FROM ./model.gguf
  1. Create the model:
ollama create tinyllama-php -f Modelfile
ollama run tinyllama-php