Back to Models
KA

kaitchup/Qwen3.6-27B-autoround-nvfp4-linearattn-BF16

kaitchupgeneral

This is Qwen/Qwen3.6-27B quantized with AutoRound to NVFP4 with linear-attention layers kept in 16-bit. The model is compatible with vLLM (tested: v0.19). Tested with an RTX Pro 6000. Currently under evaluation. Similar quantization with Qwen3.5 worked very well.

Instructions

uv pip install vllm
uv pip install git+https://github.com/huggingface/transformers.git
vllm serve [this model ID]  --max-model-len 262144 --reasoning-parser qwen3
Visit Website

0 reviews

5
0
4
0
3
0
2
0
1
0
Likes9
Downloads
📝

No reviews yet

Be the first to review kaitchup/Qwen3.6-27B-autoround-nvfp4-linearattn-BF16!

Model Info

Providerkaitchup
Categorygeneral
Reviews0
Avg. Rating / 5.0

Community

Likes9
Downloads

Rating Guidelines

★★★★★Exceptional
★★★★Great
★★★Good
★★Fair
Poor