TheBlokeAI

TheBloke's LLM work is generously supported by a grant from andreessen horowitz (a16z)


Zephyr 7B Norobots - FP16

Description

This repo contains pytorch format fp16 model files for MonsterAPI's Zephyr 7B Norobots.

It is the result of either merging a LoRA, or converting the source repository to float16.

These files were quantised using hardware kindly provided by Massed Compute.

Repositories available

Prompt template: ChatML

<|im_start|>system
{system_message}<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant

Discord

For further support, and discussions on these models and AI in general, join us at:

TheBloke AI's Discord server

Thanks, and how to contribute

Thanks to the chirper.ai team!

Thanks to Clay from gpus.llm-utils.org!

I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training.

If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects.

Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits.

Special thanks to: Aemon Algiz.

Patreon special mentions: Brandon Frisco, LangChain4j, Spiking Neurons AB, transmissions 11, Joseph William Delisle, Nitin Borwankar, Willem Michiel, Michael Dempsey, vamX, Jeffrey Morgan, zynix, jjj, Omer Bin Jawed, Sean Connelly, jinyuan sun, Jeromy Smith, Shadi, Pawan Osman, Chadd, Elijah Stavena, Illia Dulskyi, Sebastain Graf, Stephen Murray, terasurfer, Edmond Seymore, Celu Ramasamy, Mandus, Alex, biorpg, Ajan Kanaga, Clay Pascal, Raven Klaugh, 阿明, K, ya boyyy, usrbinkat, Alicia Loh, John Villwock, ReadyPlayerEmma, Chris Smitley, Cap'n Zoog, fincy, GodLy, S_X, sidney chen, Cory Kujawski, OG, Mano Prime, AzureBlack, Pieter, Kalila, Spencer Kim, Tom X Nguyen, Stanislav Ovsiannikov, Michael Levine, Andrey, Trailburnt, Vadim, Enrico Ros, Talal Aujan, Brandon Phillips, Jack West, Eugene Pentland, Michael Davis, Will Dee, webtim, Jonathan Leane, Alps Aficionado, Rooh Singh, Tiffany J. Kim, theTransient, Luke @flexchar, Elle, Caitlyn Gatomon, Ari Malik, subjectnull, Johann-Peter Hartmann, Trenton Dambrowitz, Imad Khwaja, Asp the Wyvern, Emad Mostaque, Rainer Wilmers, Alexandros Triantafyllidis, Nicholas, Pedro Madruga, SuperWojo, Harry Royden McLaughlin, James Bentley, Olakabola, David Ziegler, Ai Maven, Jeff Scroggin, Nikolai Manek, Deo Leter, Matthew Berman, Fen Risland, Ken Nordquist, Manuel Alberto Morcote, Luke Pendergrass, TL, Fred von Graf, Randy H, Dan Guido, NimbleBox.ai, Vitor Caleffi, Gabriel Tamborski, knownsqashed, Lone Striker, Erik Bjäreholt, John Detwiler, Leonard Tan, Iucharbius

Thank you to all my generous patrons and donaters!

And thank you again to a16z for their generous grant.

Original model card: MonsterAPI's Zephyr 7B Norobots

Finetuning Overview:

Model Used: HuggingFaceH4/zephyr-7b-alpha

Dataset: HuggingFaceH4/no_robots

Dataset Insights:

No Robots is a high-quality dataset of 10,000 instructions and demonstrations created by skilled human annotators. This data can be used for supervised fine-tuning (SFT) to make language models follow instructions better.

Finetuning Details:

With the utilization of MonsterAPI's LLM finetuner, this finetuning:

  • Was achieved with great cost-effectiveness.
  • Completed in a total duration of 36mins 47secs for 1 epoch using an A6000 48GB GPU.
  • Costed $1.212 for the entire epoch.

Hyperparameters & Additional Details:

  • Epochs: 1
  • Cost Per Epoch: $1.212
  • Total Finetuning Cost: $1.212
  • Model Path: HuggingFaceH4/zephyr-7b-alpha
  • Learning Rate: 0.0002
  • Data Split: 100% train
  • Gradient Accumulation Steps: 4
  • lora r: 32
  • lora alpha: 64

Prompt Structure

<|system|> <|endoftext|> <|user|> [USER PROMPT]<|endoftext|> <|assistant|> [ASSISTANT ANSWER] <|endoftext|>

Train loss :

training loss

license: apache-2.0

Downloads last month
0
Safetensors
Model size
7.24B params
Tensor type
BF16
·
Inference API
Inference API (serverless) has been turned off for this model.

Model tree for TheBloke/zephyr_7b_norobots-fp16

Adapter
(5)
this model
Quantizations
2 models

Dataset used to train TheBloke/zephyr_7b_norobots-fp16