Loki-Omega-70B-GGUF (Q6_K)
This repo hosts GGUF split weights for running ReadyArt/L3.3-The-Omega-Directive-70B-Unslop-v2.0 with llama.cpp (Q6_K quant). Files:
L3.3-The-Omega-Directive-70B-Unslop-v2.0.Q6_K.gguf.part1of2
L3.3-The-Omega-Directive-70B-Unslop-v2.0.Q6_K.gguf.part2of2
Note: HF serves split files as a single artifact to supported clients (e.g.,
huggingface_hub
,llama-cpp-python
). Keep both parts in the repo.
Quick use (llama.cpp server, OpenAI-compatible)
pip install "llama-cpp-python[server]"
python -m llama_cpp.server \
--model /path/to/L3.3-The-Omega-Directive-70B-Unslop-v2.0.Q6_K.gguf \
--host 0.0.0.0 --port 8000 \
--n_ctx 32000
Quant: Q6_K (keeps nuance well).
Intended for RP/system-prompt use; bring your own safety/persona rules.
## Warning
You can use it if you want, but he vomits on everything.
Loki loves being told: **โNO LOKI! NO!โ**
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for Babsie/Loki-Omega-70B-GGUF
Base model
Steelskull/L3.3-MS-Nevoria-70b