Text Generation
GGUF
English
gpt_oss
gpt-oss
openai
mxfp4
programming
code generation
code
coding
coder
chat
reasoning
thinking
r1
cot
deepseek
128k context
general usage
problem solving
brainstorming
solve riddles
Neo
MOE
Mixture of Experts
24 experts
NEO Imatrix
Code Imatrix
DI-Matrix
Dual Imatrix
Imatrix
imatrix
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -51,7 +51,7 @@ NEO dataset improves overall performance.
|
|
51 |
|
52 |
CODER dataset is specifically for coding performance.
|
53 |
|
54 |
-
DUEL ("DI")-> Separate Imatrix datasets (generated separately per model) are co-joined to create a new Imatrix dataset, which is then applied to the quants.
|
55 |
|
56 |
Model also passed "hard" coding test too (4 experts); no issues (IQ4_NL).
|
57 |
|
@@ -66,9 +66,9 @@ There are TWO "IQ4_NL" quants:
|
|
66 |
- OpenAI-20B-NEO-CODE-DIMAT-2-IQ4_NL.gguf : DI Imatrix applied, including output tensor (also imatrixed), and embed tensor at IQ4_NL.
|
67 |
|
68 |
There are THREE NEO MXFP4_MOE quants:
|
69 |
-
- OpenAI-20B-NEO-CODE-DIMAT-MXFP4_MOE2.gguf : Output tensor Q5_1 (
|
70 |
-
- OpenAI-20B-NEO-CODE-DIMAT-MXFP4_MOE3.gguf : Output tensor IQ4_NL (
|
71 |
-
- OpenAI-20B-NEO-CODE-DIMAT-MXFP4_MOE4.gguf : Output tensor IQ4_NL (
|
72 |
|
73 |
MXFP4_MOE quants vastly outperform (at the moment) all other quants, except IQ4_NL, Q5_1 and Q8_0 due to odd
|
74 |
issues compressing OpenAI's 20B model due to odd "tensor" dimensions.
|
|
|
51 |
|
52 |
CODER dataset is specifically for coding performance.
|
53 |
|
54 |
+
DUEL ("DI")-> Separate Imatrix datasets ("NEO" and "CODER" - generated separately per model) are co-joined to create a new Imatrix dataset, which is then applied to the quants.
|
55 |
|
56 |
Model also passed "hard" coding test too (4 experts); no issues (IQ4_NL).
|
57 |
|
|
|
66 |
- OpenAI-20B-NEO-CODE-DIMAT-2-IQ4_NL.gguf : DI Imatrix applied, including output tensor (also imatrixed), and embed tensor at IQ4_NL.
|
67 |
|
68 |
There are THREE NEO MXFP4_MOE quants:
|
69 |
+
- OpenAI-20B-NEO-CODE-DIMAT-MXFP4_MOE2.gguf : Output tensor Q5_1 (DI Imatrix applied)
|
70 |
+
- OpenAI-20B-NEO-CODE-DIMAT-MXFP4_MOE3.gguf : Output tensor IQ4_NL (DI Imatrix applied)
|
71 |
+
- OpenAI-20B-NEO-CODE-DIMAT-MXFP4_MOE4.gguf : Output tensor IQ4_NL (DI Imatrix applied) AND Embed at IQ4_NL - this makes this quant the smallest version.
|
72 |
|
73 |
MXFP4_MOE quants vastly outperform (at the moment) all other quants, except IQ4_NL, Q5_1 and Q8_0 due to odd
|
74 |
issues compressing OpenAI's 20B model due to odd "tensor" dimensions.
|