MiniMaid-L1
Introducing Our Brand New Open-sourced AI model named MiniMaid-L1, Minimaid Boast a staggering 1B params with Good Coherent Story telling, Capable roleplaying ability (Due to its 1B params, it might produce bad and repetitive output).
MiniMaid-L1 achieve a good Performance through process of DPO and Combined Heavy Finetuning, To Prevent Overfitting, We used high LR decays, And Introduced Randomization techniques to prevent the AI from learning and memorizing, However since training this on Google Colab is difficult, the Model might underperform or underfit on specific tasks Or overfit on knowledge it manage to latched on! However please be guided that we did our best, and it will improve as we move onwards!
MiniMaid-L1 is Our Smallest Model Yet! if you find any issue, then please don't hesitate to email us at:
[email protected], about any overfitting, or improvements for the future Model **C**, Once again feel free to Modify the LORA to your likings, However please consider Adding this Page for credits and if you'll increase its **Dataset**, then please handle it with care and ethical considerationsMiniMaid-L1 is
- Developed by: N-Bot-Int
- License: apache-2.0
- Parent Model from model: unsloth/llama-3.2-3b-instruct-unsloth-bnb-1bit
- Dataset Combined Using: Mosher-R1(Propietary Software)
MiniMaid-L1 Official Metric Score
Metrics Made By ItsMeDevRoland Which compares:
- Deepseek R1 3B GGUF
- Dolphin 3B GGUF
- Hermes 3b Llama GGUFF
- OpenElla3-Llama3.2B GGUFF Which are All Ranked with the Same Prompt, Same Temperature, Same Hardware(Google Colab), To Properly Showcase the differences and strength of the Models
THIS MODEL EXCELLS IN LONGER PROMPT AND STAYING IN CHARACTER BUT LAGS BEHIND DEEPSEEK-R1
Notice
- For a Good Experience, Please use
- Low temperature 1.5, min_p = 0.1 and max_new_tokens = 128
- For a Good Experience, Please use
Detail card:
Parameter
- 1 Billion Parameters
- (Please visit your GPU Vendor if you can Run 1B models)
Finetuning tool:
Unsloth AI
- This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.
- This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.
Fine-tuned Using:
Google Colab
- Downloads last month
- 0
Model tree for N-Bot-Int/MiniMaid-L1
Base model
meta-llama/Llama-3.2-1B