HumanLlama3.2-1B

This is a test model tuning 1B Llama 3.2 with ORPO:

orpo_args = ORPOConfig(
    learning_rate=8e-6,
    lr_scheduler_type="linear",
    max_length=1024,
    max_prompt_length=512,
    beta=0.1,
    per_device_train_batch_size=2,
    per_device_eval_batch_size=2,
    gradient_accumulation_steps=4,
    optim="paged_adamw_8bit",
    num_train_epochs=1,
    eval_strategy="steps",
    eval_steps=0.2,
    logging_steps=1,
    warmup_steps=10,
    report_to="wandb",
    output_dir="./results/",
)
Downloads last month
12
Safetensors
Model size
1.24B params
Tensor type
FP16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for nbeerbower/HumanLlama-3.2-1B

Finetuned
(443)
this model
Quantizations
2 models

Dataset used to train nbeerbower/HumanLlama-3.2-1B