--- library_name: transformers tags: - trl - sft license: mit --- # Model Card for Model ID ## Model Details ### Model Description This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** Sefika - **Language(s) (NLP):** EN - **License:** MIT - **Finetuned from model [optional]:** https://huggingface.co/meta-llama/Llama-2-7b-chat-hf ### Model Sources [optional] - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ### Direct Use ```python from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline tokenizer = "meta-llama/Llama-2-7b-chat-hf" model_id = "CRE_llama_fewrel_1_memory_10_4" tokenizer = AutoTokenizer.from_pretrained(model_id, use_auth_token=True) model = AutoModelForCausalLM.from_pretrained( model_id, device_map="auto", load_in_4bit=True, # Requires bitsandbytes torch_dtype="auto" ) ``` #### Testing Data FewRel **BibTeX:** The paper "Large Language Models for Continual Relation Extraction" is submitted to Springer Machine Learning journal ## Model Card Contact sefika efeoglu