metadata
license: apache-2.0
tags:
- unsloth
- Uncensored
- text-generation-inference
- transformers
- unsloth
- llama
- trl
- roleplay
- conversational
datasets:
- iamketan25/roleplay-instructions-dataset
- N-Bot-Int/Iris-Uncensored-R1
- N-Bot-Int/Moshpit-Combined-R2-Uncensored
- N-Bot-Int/Mushed-Dataset-Uncensored
- N-Bot-Int/Muncher-R1-Uncensored
- N-Bot-Int/Millia-R1_DPO
language:
- en
base_model:
- N-Bot-Int/MiniMaid-L2
pipeline_tag: text-generation
metrics:
- character
new_version: N-Bot-Int/MiniMaid-L3
Support Us Through
- [https://ko-fi.com/nexusnetworkint](Official Ko-FI link!)
GGUF Version
GGUF with Quants! Allowing you to run models using KoboldCPP and other AI Environments!
Quantizations:
Quant Type | Benefits | Cons |
---|---|---|
Q4_K_M | β Smallest size (fastest inference) | β Lowest accuracy compared to other quants |
β Requires the least VRAM/RAM | β May struggle with complex reasoning | |
β Ideal for edge devices & low-resource setups | β Can produce slightly degraded text quality | |
Q5_K_M | β Better accuracy than Q4, while still compact | β Slightly larger model size than Q4 |
β Good balance between speed and precision | β Needs a bit more VRAM than Q4 | |
β Works well on mid-range GPUs | β Still not as accurate as higher-bit models | |
Q8_0 | β Highest accuracy (closest to full model) | β Requires significantly more VRAM/RAM |
β Best for complex reasoning & detailed outputs | β Slower inference compared to Q4 & Q5 | |
β Suitable for high-end GPUs & serious workloads | β Larger file size (takes more storage) |
Model Details:
Read the Model details on huggingface Model Detail Here!