--- base_model: unsloth/mistral-7b-v0.3-bnb-4bit tags: - text-generation-inference - transformers - unsloth - mistral - gguf license: apache-2.0 language: - en --- # Uploaded model - **Developed by:** LuuWee - **License:** apache-2.0 - **Finetuned from model :** unsloth/mistral-7b-v0.3-bnb-4bit This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [](https://github.com/unslothai/unsloth) Info: I trained these Models on Google Colab with a Dataset i created out of the official CPE-Dictionary. The Dataset is formatted in the Alpaca Format: alpaca_prompt = """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request. ### Instruction: {} ### Input: {} ### Response: {}""" For the best results with this Model use this format when interacting with the model: prompt = alpaca_prompt.format(f"What is the CPE for {vendor} {productname}. Only return the CPE", "", "") this is the exact wording i used i the dataset. Input and Response should be left blank.