metadata
license: apache-2.0
tags:
- unsloth
- Uncensored
- text-generation-inference
- transformers
- unsloth
- llama
- trl
- roleplay
- conversational
datasets:
- iamketan25/roleplay-instructions-dataset
- N-Bot-Int/Iris-Uncensored-R1
- N-Bot-Int/Moshpit-Combined-R2-Uncensored
- N-Bot-Int/Mushed-Dataset-Uncensored
- N-Bot-Int/Muncher-R1-Uncensored
- N-Bot-Int/Millia-R1_DPO
language:
- en
base_model:
- N-Bot-Int/MiniMaid-L1
pipeline_tag: text-generation
metrics:
- character
new_version: N-Bot-Int/MiniMaid_L1-GGUF
Support Us Through
- [https://ko-fi.com/nexusnetworkint](Official Ko-FI link!)
MiniMaid-L1 is Depricated, and only kept for Archival purposes, and for reverting if the latest model underperforms!
GGUF Version
GGUF with Quants! Allowing you to run models using KoboldCPP and other AI Environments!
Quantizations:
Quant Type | Benefits | Cons |
---|---|---|
Q4_K_M | β Smallest size (fastest inference) | β Lowest accuracy compared to other quants |
β Requires the least VRAM/RAM | β May struggle with complex reasoning | |
β Ideal for edge devices & low-resource setups | β Can produce slightly degraded text quality | |
Q5_K_M | β Better accuracy than Q4, while still compact | β Slightly larger model size than Q4 |
β Good balance between speed and precision | β Needs a bit more VRAM than Q4 | |
β Works well on mid-range GPUs | β Still not as accurate as higher-bit models | |
Q8_0 | β Highest accuracy (closest to full model) | β Requires significantly more VRAM/RAM |
β Best for complex reasoning & detailed outputs | β Slower inference compared to Q4 & Q5 | |
β Suitable for high-end GPUs & serious workloads | β Larger file size (takes more storage) |
Model Details:
Read the Model details on huggingface Model Detail Here!