--- base_model: meta-llama/Meta-Llama-3.1-70B-Instruct library_name: peft --- # Model Card for Model ID This LoRA adapter was extracted from [mlabonne/Meta-Llama-3.1-70B-Instruct-lorablated](https://huggingface.co/mlabonne/Meta-Llama-3.1-70B-Instruct-lorablated) and uses [meta-llama/Llama-3.1-70B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-70B-Instruct) as a base. ## Model Details The model was extracted by running [mlabonne/harmful_behaviors](https://huggingface.co/datasets/mlabonne/harmful_behaviors) and the user prompts (but not assistant responses or system messages) from [Guilherme34/uncensor](https://huggingface.co/datasets/Guilherme34/uncensor) through the original abliterated model to generate a dataset of prompt/completion pairs, and was trained for 2 epochs on a 8xA100s with Axolotl using FSDP. Since the original abliterated model isn't perfect at avoiding refusals, the dataset was cleaned to remove the few refusals generated prior to training. ### Model Description - **Developed by:** @reissbaker - **Funded by:** Synthetic Lab - **License:** Apache 2.0 - **Finetuned from model:** Llama 3.1 70B Instruct ## How to Get Started with the Model Run the model with one click on [glhf.chat](https://glhf.chat). #### Training Hyperparameters * BF16 mixed-precision * 4e-4 LR * Linear LR schedule * Fused AdamW optimizer