MISHANM/Polish_text_generation_Llama3_8B_instruct
This model is specially made to work well with the Polish language. It is good at doing things like answering questions. By using advanced language tools, it gives clear and accurate answers that fit the context. This makes sure it understands the details of Polish, and its responses are trustworthy and useful in many situations.
Model Details
- Language: Polish
- Tasks: Question Answering(Polish to Polish)
- Base Model: meta-llama/Meta-Llama-3-8B-Instruct
Training Details
The model is trained on approx 52,002 instruction samples.
- GPUs: 4*AMD Radeon™ PRO V620
- Training Time: 22:57:03
Inference with Transformers
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load the fine-tuned model and tokenizer
model_path = "MISHANM/Polish_text_generation_Llama3_8B_instruct"
model = AutoModelForCausalLM.from_pretrained(model_path,device_map="auto")
tokenizer = AutoTokenizer.from_pretrained(model_path)
# Function to generate text
def generate_text(prompt, max_length=1000, temperature=0.9):
# Format the prompt according to the chat template
messages = [
{
"role": "system",
"content": "You are a Polish language expert and linguist, with same knowledge give response in Polish language.",
},
{"role": "user", "content": prompt}
]
# Apply the chat template
formatted_prompt = f"<|system|>{messages[0]['content']}<|user|>{messages[1]['content']}<|assistant|>"
# Tokenize and generate output
inputs = tokenizer(formatted_prompt, return_tensors="pt")
output = model.generate(
**inputs, max_new_tokens=max_length, temperature=temperature, do_sample=True
)
return tokenizer.decode(output[0], skip_special_tokens=True)
# Example usage
prompt = """Podaj przykład pytania, na które chatbot może odpowiedzieć."""
response_text = generate_text(prompt)
print(response_text)
Citation Information
@misc{MISHANM/Polish_text_generation_Llama3_8B_instruct,
author = {Mishan Maurya},
title = {Introducing Fine Tuned LLM for Polish Language},
year = {2025},
publisher = {Hugging Face},
journal = {Hugging Face repository},
}
- PEFT 0.12.0
- Downloads last month
- 6
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for MISHANM/Polish_text_generation_Llama3_8B_instruct
Base model
meta-llama/Meta-Llama-3-8B-Instruct