--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - mistral - trl --- # Model Card: Minimalist Assistant ## Model Details - **Base Model**: Mistral Instruct v2 - **Tokenizer**: based on Mistral Instruction following ## Intended Use - As Editor Assistant for revision and paraphrasing - Avoids technical jargon in favor of clear and accessible language ## Training Data - **Initial Training**: 14,000 conversations in minimalist style and more accessible language - Dataset: kevin009/system-defined-sft-llama3-14k - **Further Training**: 8,000 revision conversations to enhance rewriting and paraphrasing tasks. ## Performance and Limitations - **Limitations**: - May produce shorter outputs compared to original version. - Potential biases ## Ethical Considerations - Designed for daily use, potential biases from training data should be considered - The model does not have implemented safety measures to prevent generation of potentially harmful or offensive content ## Additional Information - Fine-tuned to address limitations in writing tasks observed in other models - Personalized for everyday use cases - Motivation for development was to create a model better suited for writing tasks, as existing models were found lacking in this area - SFT fine-tuned model