|
--- |
|
base_model: bleta-logjike-27b |
|
tags: |
|
- text-generation-inference |
|
- transformers |
|
- albanian |
|
- gemma3 |
|
- reasoning |
|
- mathematics |
|
- grpo |
|
- gsm8k |
|
- conversational |
|
license: apache-2.0 |
|
language: |
|
- al |
|
inference: |
|
parameters: |
|
temperature: 0.7 |
|
top_p: 0.95 |
|
top_k: 64 |
|
max_new_tokens: 512 |
|
--- |
|
|
|
# Bleta-Logjike 27B Albanian Logical Reasoning Model |
|
|
|
## Model Description |
|
- **Developed by:** klei aliaj & Armir Celiku |
|
- **Model type:** Bleta-Logjike 27B optimized for Albanian logical reasoning |
|
- **License:** apache-2.0 |
|
- **Format:** Full-precision model (HuggingFace Transformers format) |
|
- **Language:** Albanian |
|
- **Base architecture:** Based on Gemma 3 27B |
|
|
|
This model is the full-precision version of the Bleta-Logjike 27B model, specifically optimized for logical reasoning tasks in the Albanian language. Bleta is an Albanian adaptation based on Google's Gemma 3 architecture, with this version focused on enhancing logical reasoning and problem-solving capabilities for Albanian speakers. |
|
|
|
## Capabilities & Features |
|
|
|
### Logical Reasoning Focus |
|
This Albanian language model excels at: |
|
|
|
1. Logical analysis and deduction in Albanian |
|
2. Step-by-step problem solving |
|
3. Structured reasoning for complex problems |
|
4. Understanding logical relationships and dependencies |
|
5. Mathematical reasoning for grade-school level problems |
|
6. Conversational reasoning and explanations |
|
|
|
### Albanian Language Optimization |
|
- Native support for Albanian grammar and vocabulary |
|
- Understanding of Albanian cultural context |
|
- Handling of Albanian-specific logical expressions and constructs |
|
- Natural conversational abilities in Albanian |
|
|
|
## Training Methodology |
|
|
|
### GRPO Approach |
|
This model was fine-tuned using Generative Rejection Policy Optimization (GRPO), a reinforcement learning technique that trains models to optimize for specific reward functions. GRPO allows the model to learn from feedback on its generated responses, improving reasoning quality over time by: |
|
|
|
1. Generating multiple candidate responses |
|
2. Evaluating responses against specific reward criteria |
|
3. Learning to prefer high-quality reasoning patterns |
|
4. Optimizing for step-by-step problem solving |
|
|
|
### GSM8K Dataset |
|
The training utilized the GSM8K (Grade School Math 8K) dataset, which contains over 8,000 high-quality grade school math problems, requiring step-by-step reasoning to solve. The dataset provides: |
|
|
|
- Diverse mathematical problem types |
|
- Multi-step reasoning challenges |
|
- Clear step-by-step solutions |
|
- Grade-school level complexity |
|
|
|
This dataset was adapted for Albanian language training to ensure the model can handle mathematical reasoning tasks in Albanian. |
|
|
|
## Technical Specifications |
|
|
|
### Model Architecture |
|
- 27B parameters |
|
- Based on Gemma 3 architecture with Albanian adaptations |
|
- 128K context window |
|
- QK normalization |
|
- 5 sliding + 1 global attention pattern |
|
- 1024 sliding window attention |
|
|
|
### Usage Requirements |
|
- Recommended minimum 48GB GPU VRAM for full-precision inference |
|
- Compatible with Hugging Face Transformers library |
|
- Can be loaded with 4-bit or 8-bit quantization for lower resource environments |
|
|
|
## Usage with Transformers |
|
|
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
model_name = "klei1/bleta-logjike-27b" |
|
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", load_in_8bit=True) |
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
|
|
messages = [ |
|
{"role": "user", "content": "Si llogaritet sipërfaqja e një trekëndëshi?"} |
|
] |
|
|
|
text = tokenizer.apply_chat_template(messages, add_generation_prompt=True, tokenize=False) |
|
inputs = tokenizer(text, return_tensors="pt").to(model.device) |
|
|
|
outputs = model.generate(**inputs, max_new_tokens=512, temperature=0.7, top_p=0.95) |
|
print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
|
``` |
|
|
|
## Limitations |
|
|
|
This is the full-precision version of the model requiring significant computational resources. For deployment on consumer hardware, consider using the 8-bit quantized GGUF version available at klei1/bleta-logjike-27b-finetune. |
|
|
|
## Acknowledgments |
|
- Google for developing the Gemma 3 architecture |
|
- OpenAI for the GSM8K dataset |
|
- Hugging Face for their TRL library and GRPO implementation |