MiniMaid_L1-GGUF / README.md
ItsMeDevRoland's picture
Update README.md
9d8eb35 verified
metadata
license: apache-2.0
tags:
  - unsloth
  - Uncensored
  - text-generation-inference
  - transformers
  - unsloth
  - llama
  - trl
  - roleplay
  - conversational
datasets:
  - iamketan25/roleplay-instructions-dataset
  - N-Bot-Int/Iris-Uncensored-R1
  - N-Bot-Int/Moshpit-Combined-R2-Uncensored
  - N-Bot-Int/Mushed-Dataset-Uncensored
  - N-Bot-Int/Muncher-R1-Uncensored
  - N-Bot-Int/Millia-R1_DPO
language:
  - en
base_model:
  - N-Bot-Int/MiniMaid-L1
pipeline_tag: text-generation
metrics:
  - character
new_version: N-Bot-Int/MiniMaid_L1-GGUF

Support Us Through

image/png

MiniMaid-L1 is Depricated, and only kept for Archival purposes, and for reverting if the latest model underperforms!

GGUF Version

GGUF with Quants! Allowing you to run models using KoboldCPP and other AI Environments!

Quantizations:

Quant Type Benefits Cons
Q4_K_M βœ… Smallest size (fastest inference) ❌ Lowest accuracy compared to other quants
βœ… Requires the least VRAM/RAM ❌ May struggle with complex reasoning
βœ… Ideal for edge devices & low-resource setups ❌ Can produce slightly degraded text quality
Q5_K_M βœ… Better accuracy than Q4, while still compact ❌ Slightly larger model size than Q4
βœ… Good balance between speed and precision ❌ Needs a bit more VRAM than Q4
βœ… Works well on mid-range GPUs ❌ Still not as accurate as higher-bit models
Q8_0 βœ… Highest accuracy (closest to full model) ❌ Requires significantly more VRAM/RAM
βœ… Best for complex reasoning & detailed outputs ❌ Slower inference compared to Q4 & Q5
βœ… Suitable for high-end GPUs & serious workloads ❌ Larger file size (takes more storage)

Model Details:

Read the Model details on huggingface Model Detail Here!