DavidAU commited on
Commit
dd3abb7
·
verified ·
1 Parent(s): a91b924

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -51,7 +51,7 @@ NEO dataset improves overall performance.
51
 
52
  CODER dataset is specifically for coding performance.
53
 
54
- DUEL ("DI")-> Separate Imatrix datasets (generated separately per model) are co-joined to create a new Imatrix dataset, which is then applied to the quants.
55
 
56
  Model also passed "hard" coding test too (4 experts); no issues (IQ4_NL).
57
 
@@ -66,9 +66,9 @@ There are TWO "IQ4_NL" quants:
66
  - OpenAI-20B-NEO-CODE-DIMAT-2-IQ4_NL.gguf : DI Imatrix applied, including output tensor (also imatrixed), and embed tensor at IQ4_NL.
67
 
68
  There are THREE NEO MXFP4_MOE quants:
69
- - OpenAI-20B-NEO-CODE-DIMAT-MXFP4_MOE2.gguf : Output tensor Q5_1 (NEO Imatrix)
70
- - OpenAI-20B-NEO-CODE-DIMAT-MXFP4_MOE3.gguf : Output tensor IQ4_NL (NEO Imatrix)
71
- - OpenAI-20B-NEO-CODE-DIMAT-MXFP4_MOE4.gguf : Output tensor IQ4_NL (NEO Imatrix) AND Embed at IQ4_NL - this makes this quant the smallest version.
72
 
73
  MXFP4_MOE quants vastly outperform (at the moment) all other quants, except IQ4_NL, Q5_1 and Q8_0 due to odd
74
  issues compressing OpenAI's 20B model due to odd "tensor" dimensions.
 
51
 
52
  CODER dataset is specifically for coding performance.
53
 
54
+ DUEL ("DI")-> Separate Imatrix datasets ("NEO" and "CODER" - generated separately per model) are co-joined to create a new Imatrix dataset, which is then applied to the quants.
55
 
56
  Model also passed "hard" coding test too (4 experts); no issues (IQ4_NL).
57
 
 
66
  - OpenAI-20B-NEO-CODE-DIMAT-2-IQ4_NL.gguf : DI Imatrix applied, including output tensor (also imatrixed), and embed tensor at IQ4_NL.
67
 
68
  There are THREE NEO MXFP4_MOE quants:
69
+ - OpenAI-20B-NEO-CODE-DIMAT-MXFP4_MOE2.gguf : Output tensor Q5_1 (DI Imatrix applied)
70
+ - OpenAI-20B-NEO-CODE-DIMAT-MXFP4_MOE3.gguf : Output tensor IQ4_NL (DI Imatrix applied)
71
+ - OpenAI-20B-NEO-CODE-DIMAT-MXFP4_MOE4.gguf : Output tensor IQ4_NL (DI Imatrix applied) AND Embed at IQ4_NL - this makes this quant the smallest version.
72
 
73
  MXFP4_MOE quants vastly outperform (at the moment) all other quants, except IQ4_NL, Q5_1 and Q8_0 due to odd
74
  issues compressing OpenAI's 20B model due to odd "tensor" dimensions.