YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Remarks

2nd iteration has an overall worse performance running at f16 26t/s compared to the first iteration q8_0 60t/s both gguf on ollama with rtx 2070. not sure why the 2nd iteration model (f16) gives a very different output compared to the sample test output from unsloth (i presume running from lora_model safetensors). For now, v1.2 model and dataset is discontinued and will continue further iterations with the first iteration method

Downloads last month
4
GGUF
Model size
3.21B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

5-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Collection including fevohh/RayExtract-3B-v1.2