YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

airline-gpt-4o-llama-trained

This model is a fine-tuned version of meta-llama/Llama-3.3-70B-Instruct on tool-using traces.

Training details

  • Base model: meta-llama/Llama-3.3-70B-Instruct
  • Dataset: jkazdan/taubench_traces_training_data
  • Training type: LoRA fine-tuning
  • LoRA rank: 64
  • LoRA alpha: 16
  • Learning rate: 2e-05
  • Epochs: 1
  • Sequence length: 4096
  • Batch size: 1 * 16 (gradient accumulation)

This model is fine-tuned to use tools effectively based on demonstration traces.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support