FinCreditLlama-3.2-3B

๋ชจ๋ธ ๊ฐœ์š”

FinCreditLlama-3.2-3B๋Š” ๊ธˆ์œต ์‹ ์šฉ ํ‰๊ฐ€๋ฅผ ์œ„ํ•ด ํŠน๋ณ„ํžˆ ์„ค๊ณ„๋œ ํ•œ๊ตญ์–ด ์–ธ์–ด ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค.

๋ฒ ์ด์Šค ๋ชจ๋ธ: unsloth/Llama-3.2-3B-Instruct ๋ฐ์ดํ„ฐ์…‹: himedia/financial_dummy_data_v2 ํ•™์Šต ๋ฐฉ๋ฒ•: LoRA (Low-Rank Adaptation) ํ•™์Šต ์ผ์‹œ: 20250618_150212

๐Ÿ“Š ํ•™์Šต ๊ฒฐ๊ณผ

  • Final Training Loss: 0.2877
  • Final Validation Loss: 0.2917
  • Best Validation Loss: 0.2917 (step 1000)
  • Overall Improvement: 87.1%
  • Training Time: 155.65 minutes

ํ•˜์ดํผํŒŒ๋ผ๋ฏธํ„ฐ

  • Learning Rate: 0.0002
  • Max Steps: 1000
  • Batch Size: 2
  • Gradient Accumulation: 8
  • LoRA r: 64
  • LoRA alpha: 64
  • Max Sequence Length: 2048
  • Warmup Steps: 5

๐Ÿ”ง ๋ฉ”๋ชจ๋ฆฌ ์‚ฌ์šฉ๋Ÿ‰

  • GPU: Tesla T4
  • Peak Memory: 3.805 GB
  • Memory Usage: 25.8%

์‚ฌ์šฉ ๋ฐฉ๋ฒ•

from transformers import AutoTokenizer, AutoModelForCausalLM

# ๋ชจ๋ธ๊ณผ ํ† ํฌ๋‚˜์ด์ € ๋กœ๋“œ
tokenizer = AutoTokenizer.from_pretrained("himedia/fincredit-lamma3.2-3b-lr2e04-bs2-r64-steps1000-20250618_150212")
model = AutoModelForCausalLM.from_pretrained("himedia/fincredit-lamma3.2-3b-lr2e04-bs2-r64-steps1000-20250618_150212")

# ๊ฐ„๋‹จํ•œ ์ถ”๋ก  ์˜ˆ์ œ
prompt = "๊ณ ๊ฐ์˜ ์‹ ์šฉ๋“ฑ๊ธ‰์„ ํ‰๊ฐ€ํ•ด์ฃผ์„ธ์š”:"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=200)
result = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(result)

๐Ÿ“Š ํ•™์Šต ๋ฐ์ดํ„ฐ ํŒŒ์ผ

์ด ๋ ˆํฌ์ง€ํ† ๋ฆฌ์—๋Š” ๋‹ค์Œ ํ•™์Šต ๊ด€๋ จ ํŒŒ์ผ๋“ค์ด ํฌํ•จ๋˜์–ด ์žˆ์Šต๋‹ˆ๋‹ค:

  • training_log.json: ์ „์ฒด ํ•™์Šต ๋กœ๊ทธ (JSON ํ˜•์‹)
  • FinCreditLlama-3.2-3B_20250618_150212_training_curves.png: ํ•™์Šต ๊ณก์„  ์‹œ๊ฐํ™” ์ด๋ฏธ์ง€

๋ ˆํฌ์ง€ํ† ๋ฆฌ๋ช… ๊ตฌ์„ฑ

fincredit-lamma3.2-3b-lr2e04-bs2-r64-steps1000-20250618_150212 = fincredit-lamma3-4b-lr2e04-bs2-r64-steps1000-20250618_150212
  • fincredit-lamma3-4b: ๋ชจ๋ธ ๊ธฐ๋ณธ๋ช…
  • lr2e04: Learning Rate
  • bs2: Batch Size
  • r64: LoRA rank
  • steps1000: ํ•™์Šต ์Šคํ…
  • 20250618_150212: ํ•™์Šต ์‹œ๊ฐ

์„ฑ๋Šฅ

์ด ๋ชจ๋ธ์€ ํ•œ๊ตญ์–ด ๊ธˆ์œต ํ…์ŠคํŠธ์— ๋Œ€ํ•ด ํŒŒ์ธํŠœ๋‹๋˜์–ด ์‹ ์šฉ ํ‰๊ฐ€ ๊ด€๋ จ ์งˆ์˜์‘๋‹ต์— ํŠนํ™”๋˜์–ด ์žˆ์Šต๋‹ˆ๋‹ค.

๋ผ์ด์„ ์Šค

Apache 2.0

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for himedia/fincredit-lamma3.2-3b-lr2e04-bs2-r64-steps1000-20250618_150212

Finetuned
(324)
this model