File size: 4,285 Bytes
389bfca 61d2fb3 373ad66 389bfca 61d2fb3 389bfca 23c898c 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 23c898c 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 61d2fb3 373ad66 389bfca 61d2fb3 373ad66 389bfca 61d2fb3 389bfca 61d2fb3 389bfca 373ad66 389bfca 61d2fb3 389bfca 61d2fb3 373ad66 389bfca 61d2fb3 389bfca 61d2fb3 373ad66 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 |
---
library_name: transformers
tags:
- unsloth
license: cc-by-nc-2.0
base_model:
- unsloth/Meta-Llama-3.1-8B-bnb-4bit
---
# Model Card for LeoPARD 0.27
<!--MODEL STILL IN DEVELOPMENT.-->
**LeoPARD 0.27** is a fine-tuned version of **LLaMA 3.1 8B**, developed by **AxisSmart | Labs**. It incorporates **reasoning thinking** and **chain-of-thought (CoT) capabilities (beta)**, making it suitable for tasks requiring logical reasoning and step-by-step problem-solving.
---
## Model Details
### Model Description
This model is a fine-tuned version of **LLaMA 3.1 8B**, optimized for improved reasoning and chain-of-thought capabilities. It is designed to handle complex tasks that require logical thinking, structured reasoning, and multi-step problem-solving.
- **Developed by:** AxisSmart | Labs
- **Model type:** Fine-tuned language model
- **Language(s) (NLP):** Primarily English (multilingual capabilities may vary)
- **License:** [Creative Commons Attribution-NonCommercial 2.0 (CC BY-NC 2.0)]( https://creativecommons.org/licenses/by-nc/2.0/)
- **Finetuned from model:** LLaMA 3.1 8B
---
### License Details
The **CC BY-NC 4.0** license allows users to:
- **Share:** Copy and redistribute the model in any medium or format.
- **Adapt:** Remix, transform, and build upon the model for non-commercial purposes.
Under the following terms:
- **Attribution:** Users must give appropriate credit to **AxisSmart | Labs**, provide a link to the license, and indicate if changes were made.
- **NonCommercial:** The model cannot be used for commercial purposes.
For commercial use, explicit permission from **AxisSmart | Labs** is required.
---
## Uses
### Direct Use
LeoPARD 0.27 can be used directly for tasks requiring reasoning and chain-of-thought capabilities, such as:
- Logical problem-solving
- Step-by-step reasoning tasks
- Educational applications (e.g., math, science)
- Decision support systems
### Downstream Use [optional]
The model can be fine-tuned further for specific applications, such as:
- Custom reasoning pipelines
- Domain-specific problem-solving (e.g., finance, healthcare)
- Integration into larger AI systems
### Out-of-Scope Use
- Tasks requiring real-time, low-latency responses without proper optimization
- Applications involving highly sensitive or unethical use cases
- Tasks outside the scope of its reasoning and language capabilities
---
## Bias, Risks, and Limitations
- **Bias:** The model may inherit biases present in the training data or the base LLaMA model.
- **Risks:** Potential for incorrect or misleading reasoning outputs if not properly validated.
- **Limitations:** The chain-of-thought capability is still in beta and may produce incomplete or suboptimal reasoning paths.
### Recommendations
Users should validate the model's outputs, especially for critical applications. Fine-tuning on domain-specific data may improve performance and reduce biases.
---
## How to Get Started with the Model
Use the code below to load and use LeoPARD 0.27:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "model_name"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
input_text = "Explain the reasoning behind the solution to this problem: ..."
inputs = tokenizer(input_text, return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
---
## Training Details
### Training Data
The model was fine-tuned on a curated dataset designed to enhance reasoning and chain-of-thought capabilities. The dataset includes:
- Logical reasoning problems
- Step-by-step solutions
- General-purpose language data
### Training Procedure
- **Training time:** 6 hours
- **Training regime:** Mixed precision (bf16)
- **Hardware:** [Confidential]
#### Training Hyperparameters
- **Learning rate:** 2e-4
- **Batch size:** 2
- **Epochs:** 4
---
## Evaluation
Testing has not yet been conducted. Evaluation metrics and results will be added in future updates.
---
## Model Card Authors
AxisSmart | Labs
VortexHunter(Alvin)
---
## Model Card Contact
Contact Comming Soon |