As you can see, this is a LoRA model focused on reasoning in Chinese, based on LLaMA 3.2.
In my opinion, level 1 should follow the reasoning process and output results without a fixed format, avoiding chain-of-thought reasoning.
In this version, the output didn’t perform as I expected. I will work on improving it next time.
Let me know if you’d like further adjustments!
python
# Generate template
prompt = "写一首七言绝句"
reasoning_template = (
f"<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\nCutting Knowledge Date: December 2023\nToday Date: 26 July 2024\n\n<|eot_id|>"
f"<|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|>"
)
# Generate reasoning
reasoning_inputs = tokenizer(reasoning_template, return_tensors="pt").to(model.device)
reasoning_ids = model.generate(**reasoning_inputs, max_new_tokens=1024)
reasoning_output = tokenizer.decode(reasoning_ids[0, reasoning_inputs.input_ids.shape[1]:], skip_special_tokens=True)
print("|检验输出|" + reasoning_output)
Uploaded model
- Developed by: jinliuxi
- License: apache-2.0
- Finetuned from model : unsloth/llama-3.2-3b-instruct-bnb-4bit
This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API:
The model has no pipeline_tag.