π Llama-3
Collection
My experiments with Llama-3 models
β’
61 items
β’
Updated
β’
22
This model is a fine-tuned version of the powerful meta-llama/Meta-Llama-3.1-70B-Instruct
, pushing the boundaries of natural language understanding and generation even further. My goal was to create a versatile and robust model that excels across a wide range of benchmarks and real-world applications.
This model is suitable for a wide range of applications, including but not limited to:
coming soon!
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 40.30 |
IFEval (0-Shot) | 86.05 |
BBH (3-Shot) | 55.59 |
MATH Lvl 5 (4-Shot) | 21.45 |
GPQA (0-shot) | 12.53 |
MuSR (0-shot) | 17.74 |
MMLU-PRO (5-shot) | 48.48 |
This model uses ChatML
prompt template:
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
{system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>
{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
# Use a pipeline as a high-level helper
from transformers import pipeline
messages = [
{"role": "user", "content": "Who are you?"},
]
pipe = pipeline("text-generation", model="MaziyarPanahi/calme-2.3-llama3.1-70b")
pipe(messages)
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("MaziyarPanahi/calme-2.3-llama3.1-70b")
model = AutoModelForCausalLM.from_pretrained("MaziyarPanahi/calme-2.3-llama3.1-70b")
As with any large language model, users should be aware of potential biases and limitations. We recommend implementing appropriate safeguards and human oversight when deploying this model in production environments.