TriLMs unpacked to FP16 - compatible with any implementation supporting LLaMa architecture in huggingface's transformers format.
AI & ML interests
None defined yet.
Recent Activity
View all activity
QuantLMs, unpacked to FP16 format - compatible with FP16 GEMMs. After unpacking, QuantLMs have the same architecture as LLaMa.
-
SpectraSuite/QuantLM_3.9B_8bit_Unpacked
Text Generation • 4B • Updated • 12 -
SpectraSuite/QuantLM_2.3B_8bit_Unpacked
Text Generation • 2B • Updated • 16 -
SpectraSuite/QuantLM_1.5B_8bit_Unpacked
Text Generation • 2B • Updated • 11 -
SpectraSuite/QuantLM_1.1B_8bit_Unpacked
Text Generation • 1B • Updated • 13
QuantLMs, unpacked to FP16 format - compatible with FP16 GEMMs. After unpacking, QuantLMs have the same architecture as LLaMa.
-
SpectraSuite/QuantLM_3.9B_4bit_Unpacked
Text Generation • 4B • Updated • 12 -
SpectraSuite/QuantLM_2.3B_4bit_Unpacked
Text Generation • 2B • Updated • 14 -
SpectraSuite/QuantLM_1.5B_4bit_Unpacked
Text Generation • 2B • Updated • 13 -
SpectraSuite/QuantLM_1.1B_4bit_Unpacked
Text Generation • 1B • Updated • 12
FP16 LLMs with LLaMa architecture, trained on same 300B tokens as TriLMs.
QuantLMs, unpacked to FP16 format - compatible with FP16 GEMMs. After unpacking, QuantLMs have the same architecture as LLaMa.
-
SpectraSuite/QuantLM_3.9B_6bit_Unpacked
Text Generation • 4B • Updated • 11 -
SpectraSuite/QuantLM_2.3B_6bit_Unpacked
Text Generation • 2B • Updated • 14 -
SpectraSuite/QuantLM_1.5B_6bit_Unpacked
Text Generation • 2B • Updated • 14 -
SpectraSuite/QuantLM_1.1B_6bit_Unpacked
Text Generation • 1B • Updated • 30
QuantLMs, unpacked to FP16 format - compatible with FP16 GEMMs. After unpacking, QuantLMs have the same architecture as LLaMa.
-
SpectraSuite/QuantLM_3.9B_3bit_Unpacked
Text Generation • 4B • Updated • 16 -
SpectraSuite/QuantLM_2.3B_3bit_Unpacked
Text Generation • 2B • Updated • 26 -
SpectraSuite/QuantLM_1.5B_3bit_Unpacked
Text Generation • 2B • Updated • 14 -
SpectraSuite/QuantLM_1.1B_3bit_Unpacked
Text Generation • 1B • Updated • 16
TriLMs unpacked to FP16 - compatible with any implementation supporting LLaMa architecture in huggingface's transformers format.
FP16 LLMs with LLaMa architecture, trained on same 300B tokens as TriLMs.
QuantLMs, unpacked to FP16 format - compatible with FP16 GEMMs. After unpacking, QuantLMs have the same architecture as LLaMa.
-
SpectraSuite/QuantLM_3.9B_8bit_Unpacked
Text Generation • 4B • Updated • 12 -
SpectraSuite/QuantLM_2.3B_8bit_Unpacked
Text Generation • 2B • Updated • 16 -
SpectraSuite/QuantLM_1.5B_8bit_Unpacked
Text Generation • 2B • Updated • 11 -
SpectraSuite/QuantLM_1.1B_8bit_Unpacked
Text Generation • 1B • Updated • 13
QuantLMs, unpacked to FP16 format - compatible with FP16 GEMMs. After unpacking, QuantLMs have the same architecture as LLaMa.
-
SpectraSuite/QuantLM_3.9B_6bit_Unpacked
Text Generation • 4B • Updated • 11 -
SpectraSuite/QuantLM_2.3B_6bit_Unpacked
Text Generation • 2B • Updated • 14 -
SpectraSuite/QuantLM_1.5B_6bit_Unpacked
Text Generation • 2B • Updated • 14 -
SpectraSuite/QuantLM_1.1B_6bit_Unpacked
Text Generation • 1B • Updated • 30
QuantLMs, unpacked to FP16 format - compatible with FP16 GEMMs. After unpacking, QuantLMs have the same architecture as LLaMa.
-
SpectraSuite/QuantLM_3.9B_4bit_Unpacked
Text Generation • 4B • Updated • 12 -
SpectraSuite/QuantLM_2.3B_4bit_Unpacked
Text Generation • 2B • Updated • 14 -
SpectraSuite/QuantLM_1.5B_4bit_Unpacked
Text Generation • 2B • Updated • 13 -
SpectraSuite/QuantLM_1.1B_4bit_Unpacked
Text Generation • 1B • Updated • 12
QuantLMs, unpacked to FP16 format - compatible with FP16 GEMMs. After unpacking, QuantLMs have the same architecture as LLaMa.
-
SpectraSuite/QuantLM_3.9B_3bit_Unpacked
Text Generation • 4B • Updated • 16 -
SpectraSuite/QuantLM_2.3B_3bit_Unpacked
Text Generation • 2B • Updated • 26 -
SpectraSuite/QuantLM_1.5B_3bit_Unpacked
Text Generation • 2B • Updated • 14 -
SpectraSuite/QuantLM_1.1B_3bit_Unpacked
Text Generation • 1B • Updated • 16