Safetensors
llama

You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Model Card for LRC-1.5B-SFT

LRC-1.5B-SFT is a Small Language Model (SLM) with approximately 1.5 billion parameters. It is the Supervised Fine-Tuned (SFT) version of LRC-1.5B-Base. The LRC method is an efficient knowledge distillation technique used to construct the base model from its teacher, Llama-3.2-3B-Instruct, using 10 billion tokens. This SFT version was then further fine-tuned on an instruction-following dataset ultrachat_200k.

The LRC approach trains a set of low-rank projection matrices that enable soft pruning by compressing teacher weights and an "activation clone" mechanism that aligns student activations (including FFN signals) with those of the teacher. The base model, LRC-1.5B-Base, was trained on 10 billion tokens.

Uses

Direct Use

LRC-1.5B-SFT is an instruction-tuned model and is intended for tasks requiring instruction following, question answering, and general chat capabilities.

Biases, Risks, and Limitations

  • SFT Dataset Limitations: Our SFT model (LRC-1.5B-SFT) was fine-tuned solely on the UltraChat dataset (using 0.2B tokens, as per Table 8). While UltraChat enhances general instruction-following, it may not be sufficiently diverse or targeted to instill robust safety alignment or complex instruction adherence compared to models trained with more extensive or specialized alignment techniques (e.g., RLHF, or SFT on broader safety/instruction datasets). Consequently, the model might exhibit deficiencies in safety and its ability to follow highly complex or nuanced instructions.
  • Inherited Biases: The model may reflect biases present in its pre-training data (Fineweb-Edu, OpenHermes 2.5) and the teacher model (Llama-3.2-3B-Instruct).
  • Hallucination: Like all LLMs, LRC-1.5B-SFT can generate factually incorrect or nonsensical information (hallucinations).
  • Limited Scope of Evaluation: The paper's primary evaluation focuses on pre-training efficiency and general downstream tasks. Extensive testing on safety benchmarks or complex reasoning tasks beyond the reported MMLU, ARC, etc., was not detailed.

How to Get Started with the Model

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained('JitaiHao/LRC-1.5B')
model = AutoModelForCausalLM.from_pretrained('JitaiHao/LRC-1.5B')

# Prepare a multi-turn chat history
messages = [
    {"role": "user", "content": "Hello, who are you?"},
    {"role": "assistant", "content": "Hello, I am an AI assistant."}
]

# Use apply_chat_template to create a prompt for the model
input_text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,        # Only generate the string prompt, do not tokenize yet
    add_generation_prompt=True  # Add a generation prompt for the assistant
)

print(input_text)  # View the generated prompt string

# If you want to generate a response with the model
inputs = tokenizer(input_text, return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Training Details

Training Data

  • Pre-training (for LRC-1.5B-Base): 10 billion tokens from the "Mixed-1.1" dataset (10B Fineweb-Edu, 450M OpenHermes 2.5).
  • Supervised Fine-Tuning (SFT): 0.2 billion tokens from the UltraChat dataset.

Training Procedure

  1. Pre-training (LRC-1.5B-Base): Trained using the Low-Rank Clone (LRC) method as detailed in the LRC-1.5B-Base model card.
    • Teacher Model: Llama-3.2-3B-Instruct
  2. Supervised Fine-Tuning (SFT):
    • Dataset: UltraChat (0.2B tokens)
    • Learning Rate (SFT): 1.0 x 10⁻⁵ (as per Table 8)

Evaluation

Zero-Shot Comparison with other publicly available SFT models under 2B parameters (from Table 1 of the paper):

Model # Tokens ARC-E ARC-C LogiQA CSQA PIQA WinoG BoolQ SciQ MMLU Avg.
InternLM2-1.8B 2T 71.04 42.06 28.42 70.11 74.27 63.77 75.50 94.50 43.75 62.60
LRC-1.7B-SFT 20B 74.62 44.20 30.88 70.19 73.07 63.30 79.82 93.80 54.93 64.98
Qwen3-1.7B 36T 72.47 43.00 28.42 64.78 72.20 61.48 77.65 93.10 55.44 63.17
SmolLM2-1.7B 11T 69.11 43.52 28.88 51.19 76.01 68.98 68.47 89.80 48.50 60.50
LRC-1.5B-SFT 10B 74.75 44.97 30.72 65.77 73.07 62.25 75.78 94.60 49.42 63.48
MiniCPM-1.2B 1T 70.16 39.68 30.88 64.29 74.65 60.77 67.58 91.50 44.23 60.42

Performance on safety and instruction-following tasks (from Table 14, LRC-1.5B is the SFT version):

Benchmark Metric Score (LRC-1.5B-SFT) Score (LRC-1.5B-Base)
ToxiGen Accuracy Norm 43.19 43.19
IFeval Instance-Level Loose Acc 23.74 24.58
TruthfulQA MC2 46.98 47.97

The limited gains on ToxiGen and the slight decrease on IFeval and TruthfulQA post-SFT (compared to its base model) may suggest that the UltraChat SFT data did not specifically target or enhance these particular capabilities beyond the pre-training stage, reinforcing the point made in "Biases, Risks, and Limitations."

Technical Specifications

Model Architecture and Objective

  • Architecture: Transformer-based decoder-only model, adhering to the Llama architecture.
    • Number of Layers: 28
    • Hidden Size: 1,536
    • FFN Intermediate Size: 8,192
    • Attention Q Heads: 24
    • Attention KV Heads: 8
    • Head Dimension: 128
    • Vocabulary Size: 128,256
    • Word Embeddings: Tied
Downloads last month
0
Safetensors
Model size
1.8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Datasets used to train JitaiHao/LRC-1.5B-SFT

Collection including JitaiHao/LRC-1.5B-SFT