gemma-4-31B-it-The-DECKARD-HERETIC-UNCENSORED-Thinking — MLX 4.6 BPW

Mixed-precision MLX quantization of DavidAU/gemma-4-31B-it-The-DECKARD-HERETIC-UNCENSORED-Thinking, quantized with MLX Smart Quantize (MSQ) — my own sensitivity-based mixed-precision quantization method for Apple Silicon. It measures per-layer NMSE and assigns optimal bit widths automatically, combining architecture knowledge with measured data.

Details

  • Type: Vision (VLM)
  • Average: 4.6 bits per weight
  • Method: MLX Smart Quantize (MSQ)
Downloads last month
75
Safetensors
Model size
31B params
Tensor type
BF16
·
U32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for mlx-community/gemma-4-31B-it-The-DECKARD-HERETIC-UNCENSORED-Thinking-4.6bit-msq

Quantized
(162)
this model