MISHANM/Polish_text_generation_Llama3_8B_instruct

This model is specially made to work well with the Polish language. It is good at doing things like answering questions. By using advanced language tools, it gives clear and accurate answers that fit the context. This makes sure it understands the details of Polish, and its responses are trustworthy and useful in many situations.

Model Details

  1. Language: Polish
  2. Tasks: Question Answering(Polish to Polish)
  3. Base Model: meta-llama/Meta-Llama-3-8B-Instruct

Training Details

The model is trained on approx 52,002 instruction samples.

  1. GPUs: 4*AMD Radeon™ PRO V620
  2. Training Time: 22:57:03

Inference with Transformers


import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

# Load the fine-tuned model and tokenizer
model_path = "MISHANM/Polish_text_generation_Llama3_8B_instruct"

model = AutoModelForCausalLM.from_pretrained(model_path,device_map="auto")

tokenizer = AutoTokenizer.from_pretrained(model_path)

# Function to generate text
def generate_text(prompt, max_length=1000, temperature=0.9):
   # Format the prompt according to the chat template
   messages = [
       {
           "role": "system",
           "content": "You are a Polish language expert and linguist, with same knowledge give response in Polish language.",
       },
       {"role": "user", "content": prompt}
   ]

   # Apply the chat template
   formatted_prompt = f"<|system|>{messages[0]['content']}<|user|>{messages[1]['content']}<|assistant|>"

   # Tokenize and generate output
   inputs = tokenizer(formatted_prompt, return_tensors="pt")
   output = model.generate(  
       **inputs, max_new_tokens=max_length, temperature=temperature, do_sample=True
   )
   return tokenizer.decode(output[0], skip_special_tokens=True)

# Example usage
prompt = """Podaj przykład pytania, na które chatbot może odpowiedzieć."""
response_text = generate_text(prompt)
print(response_text)


Citation Information

@misc{MISHANM/Polish_text_generation_Llama3_8B_instruct,
  author = {Mishan Maurya},
  title = {Introducing Fine Tuned LLM for Polish Language},
  year = {2025},
  publisher = {Hugging Face},
  journal = {Hugging Face repository},
  
}
  • PEFT 0.12.0
Downloads last month
6
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MISHANM/Polish_text_generation_Llama3_8B_instruct

Adapter
(883)
this model

Dataset used to train MISHANM/Polish_text_generation_Llama3_8B_instruct