mega-llama-gumbo
#722
by
mangojesus
- opened
It's queued! :D
You can check for progress at http://hf.tst.eu/status.html or regularly check the model
summary page at https://hf.tst.eu/model#experimental_merge_5_1-GGUF for quants to appear.
mradermacher
changed discussion status to
closed
unfortunately, llama.cpp crashes when generating imatrix quants. this usually indicates a (value) problem with the weights, such as an internal overflov.
[ 29/ 292] blk.2.ffn_gate.weight - [ 4096, 14336, 1, 1], type = f16, converting to iq1_m .. /llmjob/llama.cpp-nocuda/ggml/src/ggml-quants.c:4453: /llmjob/llama.cpp-nocuda/ggml/src/ggml-quants.c:4453: GGML_ASSERT(besti1 >= 0 && besti2 >= 0 && best_k >= 0) failed
static quants were generated