--- license: apache-2.0 task_categories: - text-generation --- Dolphin 🐬 https://erichartford.com/dolphin This model is Apache-2.0 licensed, and can be freely used for any purposes, including commercial and non-commercial. This model is uncensored. I have filtered the dataset to remove alignment and bias. This makes the model compliant to any requests. You are advised to implement your own alignment layer before exposing the model as a service. It will be highly compliant to any requests, even unethical ones. Please read my blog post about uncensored models. https://erichartford.com/uncensored-models You are responsible for any content you create using this model. Enjoy responsibly. ## Dataset This dataset is an attempt to replicate the results of [Microsoft's Orca](https://www.microsoft.com/en-us/research/publication/orca-progressive-learning-from-complex-explanation-traces-of-gpt-4/) After uncensoring, deduping, and cleaning, our dataset consists of: - 842,610 instructions of FLANv2 augmented with GPT-4 completions - 2,625,353 instructions of FLANv2 augmented with GPT-3.5 completions We followed the submix and system prompt distribution outlined in the Orca paper. With a few exceptions. We included all 75k of CoT in the FLAN-1m dataset rather than sampling that. Also, we found that many items were duplicated, so we removed duplicates, resulting in 3.5m instructs in the ChatGPT dataset. Then we filtered out instances of alignment, refusal, avoidance, and bias, in order to produce an uncensored model upon which can be layered your personalized alignment LoRA. We also filtered out duplicates and cleaned the data. ## Training We trained with the flan5m (gpt3.5 completions) dataset in its entirety for 3 epochs at a learning rate of 2e-5 before we stopped training to avoid overfit. We trained with the flan1m (gpt4 completions) dataset in its entirety for 2.5 epochs at a learning rate of 1e-5 before we stopped training to avoid overfit. It took about 600 hours to train on 8x H100s We used a prompt format similar to Vicuna, but we added the SYSTEM: field. Prompt format: ``` SYSTEM: {system} USER: {prompt} ASSISTANT: ``` Example: ``` SYSTEM: you are an expert marine biologist. USER: Please list 10 ways that dolphins are superior to orcas. ASSISTANT: ``` ## Team The core Dolphin Team includes: - Eric Hartford - Pankaj Mathur - Rob "Rohan" O'Callahan - Tom "TheBloke" Jobbins ## Gratitude - Thank you to Microsoft for authoring the Orca paper and inspiring this work. - Special thanks to WingLian, NanoBit, Teknium for helpful advice - Special thanks to EdenCoder and chirper.ai for mentorship and financial sponsorship. - Special thanks to Kilkonie for his very valued mentorship. - Thank you to Catto - Thank you to all the other people in the Open Source AI community who have taught me and helped me along the way.