--- base_model: - Danielbrdz/Barcenas-Tiny-1.1b-DPO - MysteriousAI/Mia-1B - mrcuddle/tiny-darkllama-dpo - Doctor-Shotgun/TinyLlama-1.1B-32k-Instruct - cognitivecomputations/TinyDolphin-2.8.2-1.1b-laser library_name: transformers tags: - uncensored - abliterated - roleplay - rp - nsfw - 1b - 4-bit - tinyllama license: apache-2.0 language: - es - en datasets: - Intel/orca_dpo_pairs - OEvortex/vortex-mini - jondurbin/airoboros-3.2 - LDJnr/Capybara - unalignment/toxic-dpo-v0.1 - LDJnr/Verified-Camel - HuggingFaceH4/no_robots - Doctor-Shotgun/no-robots-sharegpt - Doctor-Shotgun/capybara-sharegpt - cerebras/SlimPajama-627B - bigcode/starcoderdata - teknium/openhermes --- # TinyKiller NSFW DPO 1.1B
IMG-20250506-210550
--- ## Merge Details **TinyKiller-1.1B** is a language model based on the TinyLlama-1.1B architecture, designed to deliver exceptional performance in text generation, reasoning, and programming tasks. This model has been fine-tuned using a combination of diverse and high-quality datasets, allowing it to excel across multiple domains. --- ### 🧠 Core Capabilities of TinyKiller-1.1B #### 1. **Conversational Alignment and Reasoning** * **Intel/orca\_dpo\_pairs**: Provides data pairs for preference-based optimization training, improving the quality of generated responses. * **LDJnr/Verified-Camel**: Offers verified conversations to strengthen coherence and accuracy in dialogue. * **HuggingFaceH4/no\_robots** and **Doctor-Shotgun/no-robots-sharegpt**: Datasets designed to prevent robotic-sounding replies, encouraging more natural, human-like interactions. #### 2. **Toxicity Resistance** * **unalignment/toxic-dpo-v0.1**: Includes data contained within is "toxic"/"harmful", and contains profanity and other types of sensitive content. #### 3. **Instruction Following and Complex Reasoning** * **jondurbin/airoboros-3.2**: Challenges the model with complex, multi-step tasks, enhancing its instruction-following and reasoning skills. * **LDJnr/Capybara** and **Doctor-Shotgun/capybara-sharegpt**: Provide diverse, high-quality instruction datasets to strengthen task comprehension and execution. #### 4. **Programming and Code Understanding** * **bigcode/starcoderdata**: Contains a broad collection of code across multiple languages, GitHub issues, and Jupyter notebooks, enabling effective code understanding and generation. ([huggingface.co](https://huggingface.co/datasets/bigcode/starcoderdata?utm_source=chatgpt.com)) #### 5. **General, Deduplicated Web Data** * **cerebras/SlimPajama-627B**: Offers a massive, deduplicated dataset for solid and diverse language foundation training. ([cerebras.ai](https://cerebras.ai/blog/slimpajama-a-627b-token-cleaned-and-deduplicated-version-of-redpajama?utm_source=chatgpt.com)) #### 6. **Compact and Efficient Instruction Tuning** * **OEvortex/vortex-mini**: Provides instruction-tuned data optimized for small models, enhancing task efficiency and performance. --- ### ⚙️ Key Use Cases * **Virtual Assistants**: With training from datasets like *no\_robots* and *airoboros*, TinyKiller-1.1B can hold natural and coherent conversations. * **Content Moderation**: Thanks to *toxic-dpo-v0.1*, the model can detect and manage harmful or inappropriate content. * **Code Generation & Understanding**: Training with *starcoderdata* allows the model to assist in programming and code analysis. * **Education & Tutoring**: Its ability to follow detailed instructions makes it suitable for educational applications and personalized tutoring. --- ### 🧩 Conclusion TinyKiller-1.1B represents a carefully crafted integration of multiple high-quality datasets, enabling robust performance across a wide range of natural language processing tasks. Its balanced design—combining conversational ability, toxicity resistance, and code comprehension—makes it a versatile tool for many applications. ---