we fine-tune this model based on mistral-7b-v0.1
Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("devhyun88/hyun-mistral-7b-orca-platypus-refine") model = AutoModelForCausalLM.from_pretrained("devhyun88/hyun-mistral-7b-orca-platypus-refine")
- Downloads last month
- 4
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support