Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
vito95311
/
Qwen3-Omni-30B-A3B-Thinking-GGUF-INT8FP16
like
13
Text Generation
GGUF
PyTorch
Transformers
Chinese
English
multilingual
llama.cpp
multimodal
quantized
ollama
llama-cpp
qwen
omni
int8
fp16
Eval Results
License:
apache-2.0
Model card
Files
Files and versions
xet
Community
7
Deploy
Use this model
main
Qwen3-Omni-30B-A3B-Thinking-GGUF-INT8FP16
65.4 GB
1 contributor
History:
6 commits
vito95311
Update README.md
10d1e93
verified
29 days ago
.gitattributes
Safe
1.56 kB
Merge with remote repository and update GGUF documentation
about 1 month ago
MODEL_CARD.md
7.38 kB
Initial GGUF release: Qwen3-Omni quantized models with Ollama support
about 1 month ago
Qwen3OmniQuantized.modelfile
453 Bytes
Initial GGUF release: Qwen3-Omni quantized models with Ollama support
about 1 month ago
README.md
11.2 kB
Update README.md
29 days ago
example_usage.py
10.5 kB
Initial GGUF release: Qwen3-Omni quantized models with Ollama support
about 1 month ago
qwen3_omni_f16.gguf
32.7 GB
xet
Initial GGUF release: Qwen3-Omni quantized models with Ollama support
about 1 month ago
qwen3_omni_quantized.gguf
32.7 GB
xet
Initial GGUF release: Qwen3-Omni quantized models with Ollama support
about 1 month ago