YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
airline-gpt-4o-llama-trained
This model is a fine-tuned version of meta-llama/Llama-3.3-70B-Instruct on tool-using traces.
Training details
- Base model: meta-llama/Llama-3.3-70B-Instruct
- Dataset: jkazdan/taubench_traces_training_data
- Training type: LoRA fine-tuning
- LoRA rank: 64
- LoRA alpha: 16
- Learning rate: 2e-05
- Epochs: 1
- Sequence length: 4096
- Batch size: 1 * 16 (gradient accumulation)
This model is fine-tuned to use tools effectively based on demonstration traces.
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support