Ollama

Modelfile

FROM Llama-3-Open-Ko-8B-Q8_0.gguf

TEMPLATE """{{- if .System }}
<s>{{ .System }}</s>
{{- end }}
<s>Human:
{{ .Prompt }}</s>
<s>Assistant:
"""

SYSTEM """A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions."""

PARAMETER temperature 0
PARAMETER num_predict 3000
PARAMETER num_ctx 4096
PARAMETER stop <s>
PARAMETER stop </s>

Update @ 2024.04.24: Release Llama-3-Open-Ko-8B model & Llama-3-Open-Ko-8B-Instruct-preview

Model Details

Llama-3-Open-Ko-8B

The Llama-3-Open-Ko-8B model is a continued pretrained language model based on the Llama-3-8B framework. This model is trained with over 60GB of deduplicated texts sourced from publicly available resources. With the new Llama-3 tokenizer, the model has been pretrained with more than 17.7B tokens, which is slightly more than that processed by the Korean tokenizer of Llama-2. Training was conducted on a TPUv5e-256, supported by Google's TRC program.

Llama-3-Open-Ko-8B-Instruct-preview

The Instruction model, named Llama-3-Open-Ko-8B-Instruct-preview, incorporates concepts from the Chat Vector paper. This model is a preview and has not been fine-tuned with any Korean instruction set, making it a strong starting point for developing new chat and instruct models.

Meta Llama-3

Developed and released by Meta, the Meta Llama 3 family of large language models (LLMs) are optimized for dialogue use cases and excel across common industry benchmarks, emphasizing helpfulness and safety.

Model Developers: Junbum Lee (Beomi)

Variations: Llama-3-Open-Ko is available in one configuration โ€” 8B.

Input/Output: Models accept text input and generate text and code.

Model Architecture: Llama 3 utilizes an optimized transformer architecture.

Training Data Params Context length GQA Token count Knowledge cutoff
Llama-3-Open-Ko Same as Open-Solar-Ko Dataset 8B 8k Yes 17.7B+ Jun, 2023

*Dataset list available here

Intended Use

Commercial and Research Applications: Llama 3 is designed for use in English, tailored for assistant-like chat in its instruction-tuned models, while the pretrained models are versatile across various natural language generation tasks.

Out-of-scope: Any use violating applicable laws, regulations, or the Acceptable Use Policy and Llama 3 Community License is prohibited.

Responsibility & Safety

Meta's commitment to Responsible AI includes steps to limit misuse and harm while supporting the open source community. Developers are encouraged to implement safety best practices and use resources like Meta Llama Guard 2 and Code Shield to tailor safety needs specifically to their use cases.

Responsible Release

Following a rigorous process against misuse, we ensure all safety and ethical guidelines are adhered to, as detailed in our Responsible Use Guide.

Ethical Considerations and Limitations

Llama 3 is built on the principles of openness, inclusivity, and helpfulness, designed to be accessible and valuable across diverse backgrounds and use cases. Developers should undertake thorough safety testing and tuning for specific applications before deployment.

Citation instructions

Llama-3-Open-Ko

@article{llama3openko,
  title={Llama-3-Open-Ko},
  author={L, Junbum},
  year={2024},
  url={https://huggingface.co/beomi/Llama-3-Open-Ko-8B}
}

Original Llama-3

@article{llama3modelcard,
  title={Llama 3 Model Card},
  author={AI@Meta},
  year={2024},
  url = {https://github.com/meta-llama/llama3/blob/main/MODEL_CARD.md}
}
Downloads last month
2,203
GGUF
Model size
8.03B params
Architecture
llama

4-bit

5-bit

6-bit

8-bit

Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.