--- license: other license_name: exaone license_link: LICENSE datasets: - huggingface-KREW/korean-role-playing language: - ko base_model: - LGAI-EXAONE/EXAONE-3.5-2.4B-Instruct tags: - exaone - role-playing --- # EXAGIRL-2.4B-Instruct EXAGIRL Logo ## πŸ‘€ λͺ¨λΈ μ†Œκ°œ **EXAGIRL-2.4B-Instruct**λŠ” LG AI Research의 [EXAONE-3.5-2.4B-Instruct](https://huggingface.co/LGAI-EXAONE/EXAONE-3.5-2.4B-Instruct) λͺ¨λΈμ„ 기반으둜, **ν•œκ΅­μ–΄ λ‘€ν”Œλ ˆμž‰ λŒ€ν™”**에 νŠΉν™”λ˜λ„λ‘ λ―Έμ„Έμ‘°μ •ν•œ LoRA 기반 λͺ¨λΈμž…λ‹ˆλ‹€. 우리의 친ꡬ '엑사'λŠ” λ‹€μ–‘ν•œ 페λ₯΄μ†Œλ‚˜μ™€ 상황에 맞좰 μžμ—°μŠ€λŸ½κ³  μž¬λ―ΈμžˆλŠ” λŒ€ν™”λ₯Ό ν•  수 μžˆλ„λ‘ λ§Œλ“€μ–΄μ‘ŒμŠ΅λ‹ˆλ‹€. > 아직 ν•™μŠ΅μ΄ μ œλŒ€λ‘œ λ˜μ§€ μ•Šμ€ pre-release μƒνƒœμž…λ‹ˆλ‹€ ## πŸ§ͺ ν•™μŠ΅ 정보 - **Base model**: [LGAI-EXAONE/EXAONE-3.5-2.4B-Instruct](https://huggingface.co/LGAI-EXAONE/EXAONE-3.5-2.4B-Instruct) - **Training strategy**: LoRA (Low-Rank Adaptation) - **learning rate**: 4e-5 - **Dataset**: - [`gf-persona-data`](https://huggingface.co/datasets/huggingface-KREW/korean-role-playing) (1 epoch) - [`exa-data`](https://huggingface.co/datasets/huggingface-KREW/korean-role-playing) (2 epoch) - **지원 μ–Έμ–΄**: ν•œκ΅­μ–΄ μ „μš© ## πŸ—ƒοΈ 데이터셋 μ„€λͺ… 이 λͺ¨λΈμ€ [huggingface-KREW/korean-role-playing](https://huggingface.co/datasets/huggingface-KREW/korean-role-playing)에 ν¬ν•¨λœ λ‹€μŒμ˜ μ„œλΈŒμ…‹μœΌλ‘œ ν•™μŠ΅λ˜μ—ˆμŠ΅λ‹ˆλ‹€: - [`gf-persona-data`](https://github.com/Pseudo-Lab/Hugging-Face-Hub-Garden/discussions/31): 연인 κ°„μ˜ 페λ₯΄μ†Œλ‚˜ 기반 μ—­ν• κ·Ή λŒ€ν™” λ°μ΄ν„°μ…‹μž…λ‹ˆλ‹€. - [`exa-data`](https://github.com/Pseudo-Lab/Hugging-Face-Hub-Garden/discussions/30): 엑사(EXA) 세계관 기반 페λ₯΄μ†Œλ‚˜λ₯Ό κ°€μ§„ μΊλ¦­ν„°μ˜ 감정 ν‘œν˜„κ³Ό 행동 μ§€μ‹œ λŒ€ν™”κ°€ ν¬ν•¨λ˜μ–΄ μžˆμŠ΅λ‹ˆλ‹€. ## πŸ§‘β€πŸ’» μ‚¬μš© μ˜ˆμ‹œ ```python from transformers import AutoTokenizer, AutoModelForCausalLM model_name = "huggingface-KREW/EXAGIRL-2.4B-Instruct" # Tokenizer & Model 뢈러였기 tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True) model = AutoModelForCausalLM.from_pretrained( model_name, torch_dtype="auto", # λ˜λŠ” torch.bfloat16 trust_remote_code=True, device_map="auto" ) # λŒ€ν™” ν”„λ‘¬ν”„νŠΈ ꡬ성 messages = [ {"role": "user", "content": "엑사야 λ­ν•˜κ³ μžˆλ‹ˆ?"} ] # Chat ν…œν”Œλ¦Ώ 적용 input_ids = tokenizer.apply_chat_template( messages, tokenize=True, add_generation_prompt=True, return_tensors="pt" ) # 응닡 생성 output = model.generate( input_ids.to(model.device), eos_token_id=tokenizer.eos_token_id, max_new_tokens=128, do_sample=False ) print(tokenizer.decode(output[0], skip_special_tokens=True)) ``` ## πŸ“œ λΌμ΄μ„ μŠ€ λ³Έ λͺ¨λΈμ€ [EXAONE AI Model License Agreement 1.1 - NC](https://huggingface.co/LGAI-EXAONE/EXAONE-3.5-2.4B-Instruct/blob/main/LICENSE)λ₯Ό λ”°λ₯΄λ©°, **비상업적 연ꡬ λͺ©μ **으둜만 μ‚¬μš©μ΄ κ°€λŠ₯ν•©λ‹ˆλ‹€. ## πŸ™Œ κΈ°μ—¬μž - [@lastdefiance20](https://huggingface.co/lastdefiance20)