Image-Text-to-Text
Safetensors
gemma4
heretic
abliterated
uncensored
finetune
unsloth
all use cases
coder
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prosing
vivid writing
fiction
roleplaying
bfloat16
conversational
4-bit precision
gemma-4-31B-it-The-DECKARD-HERETIC-UNCENSORED-Thinking — MLX 4.6 BPW
Mixed-precision MLX quantization of DavidAU/gemma-4-31B-it-The-DECKARD-HERETIC-UNCENSORED-Thinking, quantized with MLX Smart Quantize (MSQ) — my own sensitivity-based mixed-precision quantization method for Apple Silicon. It measures per-layer NMSE and assigns optimal bit widths automatically, combining architecture knowledge with measured data.
Details
- Type: Vision (VLM)
- Average: 4.6 bits per weight
- Method: MLX Smart Quantize (MSQ)
- Downloads last month
- 75
Model tree for mlx-community/gemma-4-31B-it-The-DECKARD-HERETIC-UNCENSORED-Thinking-4.6bit-msq
Base model
google/gemma-4-31B-it