Back to Models
AB

Abiray/Huihui-Qwen3.6-35B-A3B-abliterated-GGUF

Abiraygeneral

Huihui Qwen3.6-35B A3B Abliterated (GGUF)

This repository provides GGUF format quantizations for the huihui-ai/Huihui-Qwen3.6-35B-A3B-abliterated model.

Because this model has been fully "abliterated" to bypass alignment and safety refusals, it acts as a highly capable engine for unrestricted creative writing, dynamic storytelling, and immersive roleplay scenarios.

Available Quantizations

FileBit SizeDescription
huihui-35B-Q8_0.gguf8-bitHighest quality quant, virtually indistinguishable from F16.
huihui-35B-Q6_K.gguf6-bitExcellent quality with a noticeably reduced memory footprint.
huihui-35B-Q5_K_M.gguf5-bitGreat balance between reasoning performance and RAM usage.
huihui-35B-Q4_K_M.gguf4-bitRecommended. The optimal sweet spot for speed and quality.
huihui-35B-Q4_K_S.gguf4-bitSlightly smaller than K_M, allowing for faster inference on constrained setups.
huihui-35B-Q3_K_M.gguf3-bitLowest resource requirement, though perplexity loss becomes more noticeable.

Quick Start (llama.cpp)

These models are designed to be run directly via llama.cpp. The following commands are standard for local Linux environments (such as Linux Mint or Ubuntu).

1. Clone and compile via CMake:

git clone [https://github.com/ggerganov/llama.cpp](https://github.com/ggerganov/llama.cpp)
cd llama.cpp
cmake -B build
cmake --build build --config Release
Visit Website

0 reviews

5
0
4
0
3
0
2
0
1
0
Likes13
Downloads
📝

No reviews yet

Be the first to review Abiray/Huihui-Qwen3.6-35B-A3B-abliterated-GGUF!

Model Info

ProviderAbiray
Categorygeneral
Reviews0
Avg. Rating / 5.0

Community

Likes13
Downloads

Rating Guidelines

★★★★★Exceptional
★★★★Great
★★★Good
★★Fair
Poor