Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
vito95311
/
Qwen3-Omni-30B-A3B-Thinking-GGUF-INT8FP16
like
14
Text Generation
GGUF
PyTorch
Transformers
Chinese
English
multilingual
llama.cpp
multimodal
quantized
ollama
llama-cpp
qwen
omni
int8
fp16
Eval Results
License:
apache-2.0
Model card
Files
Files and versions
xet
Community
7
Deploy
Use this model
d4ef36e
Qwen3-Omni-30B-A3B-Thinking-GGUF-INT8FP16
65.4 GB
1 contributor
History:
1 commit
vito95311
Initial GGUF release: Qwen3-Omni quantized models with Ollama support
d4ef36e
3 months ago
.gitattributes
Safe
135 Bytes
Initial GGUF release: Qwen3-Omni quantized models with Ollama support
3 months ago
MODEL_CARD.md
7.38 kB
Initial GGUF release: Qwen3-Omni quantized models with Ollama support
3 months ago
Qwen3OmniQuantized.modelfile
453 Bytes
Initial GGUF release: Qwen3-Omni quantized models with Ollama support
3 months ago
README.md
9.61 kB
Initial GGUF release: Qwen3-Omni quantized models with Ollama support
3 months ago
example_usage.py
10.5 kB
Initial GGUF release: Qwen3-Omni quantized models with Ollama support
3 months ago
qwen3_omni_f16.gguf
32.7 GB
xet
Initial GGUF release: Qwen3-Omni quantized models with Ollama support
3 months ago
qwen3_omni_quantized.gguf
32.7 GB
xet
Initial GGUF release: Qwen3-Omni quantized models with Ollama support
3 months ago