QuantFactory Banner

QuantFactory/TwinLlama-3.1-8B-GGUF

This is quantized version of mlabonne/TwinLlama-3.1-8B created using llama.cpp

Original Model Card

image/png

πŸ‘₯ TwinLlama-3.1-8B

TwinLlama-3.1-8B is a model created for the LLM Engineer's Handbook, trained on mlabonne/llmtwin.

It is designed to act as a digital twin, which is a clone of myself and my co-authors (Paul Iusztin and Alex Vesa), imitating our writing style and drawing knowledge from our articles.


This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.

Downloads last month
226
GGUF
Model size
8.03B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no pipeline_tag.

Model tree for QuantFactory/TwinLlama-3.1-8B-GGUF

Quantized
(197)
this model

Dataset used to train QuantFactory/TwinLlama-3.1-8B-GGUF