Back to Models
SP

sphaela/Qwen3.6-27B-AutoRound-GGUF

sphaelageneral

Qwen3.6-27B GGUF (AutoRound Quantized)

This repository contains GGUF quantized versions of Qwen/Qwen3.6-27B created using Intel's AutoRound quantization method.

Quantization Details

The models were quantized using various schemes provided by the auto-round tool. For better compatibility and smaller size, we provide unified multimodal projector (mmproj) files in F16, BF16, and F32 formats.

Files and Sizes

File NameQuant TypeSizeDescription
Qwen3.6-27B-Q2_K_S.ggufQ2_K_S8.9 GBExtremely high compression, significant quality loss.
Qwen3.6-27B-Q2_K_MIXED.ggufQ2_K_MIXED16 GBRecommended high-compression option. Uses Q4 for KV cache with good quality. Fast inference.
Qwen3.6-27B-Q3_K_S.ggufQ3_K_S12 GBVery high compression, notable quality loss.
Qwen3.6-27B-Q3_K_M.ggufQ3_K_M12 GBBalanced 3-bit quantization.
Qwen3.6-27B-Q3_K_L.ggufQ3_K_L12 GBHigh quality 3-bit quantization.
Qwen3.6-27B-Q4_0.ggufQ4_015 GBStandard 4-bit quantization, good balance.
Qwen3.6-27B-Q4_1.ggufQ4_116 GBHigher quality 4-bit quantization than Q4_0.
Qwen3.6-27B-Q4_K_S.ggufQ4_K_S15 GBSmall 4-bit K-quant, good efficiency.
Qwen3.6-27B-Q4_K_M.ggufQ4_K_M15 GBRecommended 4-bit K-quant, excellent balance.
Qwen3.6-27B-Q5_0.ggufQ5_018 GBStandard 5-bit quantization, very high quality.
Qwen3.6-27B-Q5_1.ggufQ5_119 GBHigher quality 5-bit quantization than Q5_0.
Qwen3.6-27B-Q5_K_S.ggufQ5_K_S18 GBSmall 5-bit K-quant, very high quality.
Qwen3.6-27B-Q5_K_M.ggufQ5_K_M18 GBRecommended 5-bit K-quant, near-lossless.
Qwen3.6-27B-Q6_K.ggufQ6_K21 GB6-bit K-quant, virtually indistinguishable from F16.
Qwen3.6-27B-Q8_0.ggufQ8_029 GB8-bit quantization, near-lossless.
mmproj-model-f16.ggufF16885 MBUnified Projector in Float16 format.
mmproj-model-bf16.ggufBF16889 MBUnified Projector in BFloat16 format.
mmproj-model-f32.ggufF321.8 GBUnified Projector in Float32 format.

Generate the Model

The models were generated using Intel's AutoRound with the following command:

auto-round --model Qwen/Qwen3.6-27B --output_dir ./quantized/ --scheme <SCHEME> --iters 0

Usage with llama.cpp

These models can be used with llama.cpp. For multimodal usage, you must specify the projector file:

./llama-cli -m Qwen3.6-27B-Q4_K_M.gguf --mmproj mmproj-model-f16.gguf --image your_image.jpg -p "Describe this image."

About AutoRound

AutoRound is an advanced quantization technique from Intel that aims to minimize accuracy loss through automated rounding optimization.

Visit Website

0 reviews

5
0
4
0
3
0
2
0
1
0
Likes4
Downloads
📝

No reviews yet

Be the first to review sphaela/Qwen3.6-27B-AutoRound-GGUF!

Model Info

Providersphaela
Categorygeneral
Reviews0
Avg. Rating / 5.0

Community

Likes4
Downloads

Rating Guidelines

★★★★★Exceptional
★★★★Great
★★★Good
★★Fair
Poor