Quantized Qwen3-VL
Collection
4 items
•
Updated
This is Qwen/Qwen3-VL-2B-Instruct quantized with AutoRound in NVFP4 (llm-compressor format). The model has been created, tested, and evaluated by The Kaitchup. The model is NOT compatible with vLLM (as of v0.11).
Subscribe to The Kaitchup. This helps me a lot to continue quantizing and evaluating models for free. Or you prefer to give some GPU hours, "buy me a coffee"
Base model
Qwen/Qwen3-VL-2B-Instruct