Qwen3-Omni-30B-A3B-Instruct GGUF

GGUF weights for Qwen3-Omni-30B-A3B-Instruct in f16, q8_0, and q4_0 quantizations.

Quantization Total Size
f16 64.7 GB
q8_0 35.4 GB
q4_0 19.7 GB

The reference/ directory contains test fixtures for ./test-e2e.sh.

Downloads last month
1,069
GGUF
Model size
0.6B params
Architecture
audio-encoder
Hardware compatibility
Log In to add your hardware

4-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for DeepMentor/qwen3-omni-30B-a3b-instruct-gguf

Quantized
(15)
this model