--- license: cc-by-nc-sa-4.0 datasets: - wi_locness - matejklemen/falko_merlin - paws - paws-x - asset language: - en - de - es - ar - ja - ko - zh metrics: - bleu - rouge - sari - accuracy library_name: transformers widget: - text: 'Umschreiben sie den satz: When I grow up, I start to understand what he said is quite right.' example_title: GEC (de|en) - text: '문장의 간단한 버전 작성: Cuando se pueden mantener tasas de flujo comparables, los resultados son altos.' example_title: Simplification (ko|es) - text: 'Paraphrase this: いちごは物語を紹介し、読者をイベントに導くと彼は言った。' example_title: Paraphrase (en|ja) pipeline_tag: text2text-generation base_model: grammarly/medit-xl tags: - llama-cpp - gguf-my-lora --- # aynig/medit-xl-F16-GGUF This LoRA adapter was converted to GGUF format from [`grammarly/medit-xl`](https://huggingface.co/grammarly/medit-xl) via the ggml.ai's [GGUF-my-lora](https://huggingface.co/spaces/ggml-org/gguf-my-lora) space. Refer to the [original adapter repository](https://huggingface.co/grammarly/medit-xl) for more details. ## Use with llama.cpp ```bash # with cli llama-cli -m base_model.gguf --lora medit-xl-f16.gguf (...other args) # with server llama-server -m base_model.gguf --lora medit-xl-f16.gguf (...other args) ``` To know more about LoRA usage with llama.cpp server, refer to the [llama.cpp server documentation](https://github.com/ggerganov/llama.cpp/blob/master/examples/server/README.md).