|
--- |
|
license: apache-2.0 |
|
tags: |
|
- merge |
|
- mergekit |
|
- lazymergekit |
|
- akameswa/mistral-7b-instruct-javascript-16bit |
|
- akameswa/mistral-7b-instruct-python-16bit |
|
- akameswa/mistral-7b-instruct-java-16bit |
|
- akameswa/mistral-7b-instruct-javascript-16bit |
|
--- |
|
|
|
# mistral-7b-instruct-code-ties |
|
|
|
mistral-7b-instruct-code-ties is a merge of the following models using [mergekit](https://github.com/cg123/mergekit): |
|
* [akameswa/mistral-7b-instruct-javascript-16bit](https://huggingface.co/akameswa/mistral-7b-instruct-javascript-16bit) |
|
* [akameswa/mistral-7b-instruct-python-16bit](https://huggingface.co/akameswa/mistral-7b-instruct-python-16bit) |
|
* [akameswa/mistral-7b-instruct-java-16bit](https://huggingface.co/akameswa/mistral-7b-instruct-java-16bit) |
|
* [akameswa/mistral-7b-instruct-javascript-16bit](https://huggingface.co/akameswa/mistral-7b-instruct-javascript-16bit) |
|
|
|
## 🧩 Configuration |
|
|
|
```yaml |
|
models: |
|
- model: akameswa/mistral-7b-instruct-v0.2-bnb-16bit |
|
- model: akameswa/mistral-7b-instruct-javascript-16bit |
|
parameters: |
|
density: 0.85 |
|
weight: 0.25 |
|
- model: akameswa/mistral-7b-instruct-python-16bit |
|
parameters: |
|
density: 0.85 |
|
weight: 0.25 |
|
- model: akameswa/mistral-7b-instruct-java-16bit |
|
parameters: |
|
density: 0.85 |
|
weight: 0.25 |
|
- model: akameswa/mistral-7b-instruct-javascript-16bit |
|
parameters: |
|
density: 0.85 |
|
weight: 0.25 |
|
merge_method: ties |
|
base_model: akameswa/mistral-7b-instruct-v0.2-bnb-16bit |
|
parameters: |
|
normalize: true |
|
dtype: float16 |
|
``` |
|
|
|
## Inference |
|
```python |
|
from unsloth import FastLanguageModel |
|
import torch |
|
|
|
model, tokenizer = FastLanguageModel.from_pretrained( |
|
model_name = "akameswa/mistral-7b-instruct-code-ties", |
|
max_seq_length = 2048, |
|
) |
|
|
|
xlcost_prompt = """Below is a description of a programming task. Write a response that appropriately completes the task based on the given description. |
|
|
|
### Description: |
|
{} |
|
|
|
### Code: |
|
{}""" |
|
|
|
FastLanguageModel.for_inference(model) |
|
inputs = tokenizer( |
|
[ |
|
xlcost_prompt.format( |
|
"Continue the fibonnaci sequence.", |
|
"", |
|
) |
|
], return_tensors = "pt").to("cuda") |
|
|
|
outputs = model.generate(**inputs, max_new_tokens = 64, use_cache = True) |
|
tokenizer.batch_decode(outputs) |
|
``` |
|
|
|
* [Link to inference notebook](https://github.com/akameswa/CodeGenerationMoE/blob/main/code/inference_merge.ipynb) |