mega-llama-gumbo

#722
by mangojesus - opened

It's queued! :D

You can check for progress at http://hf.tst.eu/status.html or regularly check the model
summary page at https://hf.tst.eu/model#experimental_merge_5_1-GGUF for quants to appear.

mradermacher changed discussion status to closed

unfortunately, llama.cpp crashes when generating imatrix quants. this usually indicates a (value) problem with the weights, such as an internal overflov.

[ 29/ 292] blk.2.ffn_gate.weight - [ 4096, 14336, 1, 1], type = f16, converting to iq1_m .. /llmjob/llama.cpp-nocuda/ggml/src/ggml-quants.c:4453: /llmjob/llama.cpp-nocuda/ggml/src/ggml-quants.c:4453: GGML_ASSERT(besti1 >= 0 && besti2 >= 0 && best_k >= 0) failed

static quants were generated

Sign up or log in to comment