🏜️MIRAGE-Bench [NAACL'25]
Collection
Dataset Collection from the MIRAGE-Bench paper
•
13 items
•
Updated
•
1
This model is a fine-tuned version of meta-llama/Meta-Llama-3-8B-Instruct on the nthakur/mirage-bench-sft-teacher-gpt-4o dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
0.3599 | 0.1268 | 200 | 0.3172 |
0.3295 | 0.2536 | 400 | 0.2919 |
0.323 | 0.3803 | 600 | 0.2789 |
0.3274 | 0.5071 | 800 | 0.2686 |
0.3171 | 0.6339 | 1000 | 0.2597 |
0.3034 | 0.7607 | 1200 | 0.2540 |
0.265 | 0.8875 | 1400 | 0.2510 |
Base model
meta-llama/Meta-Llama-3-8B-Instruct