What happened to gemma-4 mxfp8 quanting?
You might have noticed my numbers changing lately a few times. I finally found why. The quants I published for the E4B were made on April 8.
With the updated template, the numbers as currently displayed, stand
However, I re-quanted some of the E4B because I had removed them from the local repo, and the ones quanted yesterday are different. Not better
gemma-4-E4B-it-The-DECKARD-Expresso-Universe-HERETIC-UNCENSORED
quant arc arc/e boolq hswag obkqa piqa wino
mxfp8 0.518,0.709,0.755,0.657,0.418,0.759,0.626
Quant created yesterday:
mxfp8 0.508,0.707,0.756,0.658,0.424,0.760,0.624
This is missing from the tensors:
language_model.model.per_layer_model_projection.scales
also missing in the recent mxfp8 of gemma-4-26B-A4B-it:
language_model.model.layers.[0-29].mlp.down_proj.biases
language_model.model.layers.[0-29].mlp.gate_proj.biases
language_model.model.layers.[0-29].mlp.up_proj.biases
language_model.model.layers.[0-29].router.proj.biases