Back to Models
TH

TheCluster/Darwin-35B-A3B-Opus-MLX-mxfp8

TheClusterimage

Darwin-35B-A3B-Opus

Quality: quantized (mxfp8, 8.349 bpw)

Model Specifications

ArchitectureQwen3.5 MoE (Gated DeltaNet + MoE)
Total Parameters35B
Active Parameters3B per forward pass
Layers40
Layout10 x (3 x GDN-MoE + 1 x Attention-MoE)
Experts256 (8 routed + 1 shared active)
Context Length262,144 native
Languages201
MultimodalImage and Video
LicenseApache 2.0

Parent Models

Both parents share the identical Qwen3.5-35B-A3B architecture (40 layers, 256 experts, GDN+MoE hybrid). The Mother is a LoRA SFT on the same base — not a different architecture. "Text-only" refers to the training data (Claude 4.6 Opus reasoning chains), not the model structure.

RoleModelArchitectureTraining
FatherQwen/Qwen3.5-35B-A3BQwen3.5-35B-A3BOriginal pre-training + RLHF
MotherJackrong/Qwen3.5-35B-A3B-Claude-4.6-Opus-Reasoning-DistilledQwen3.5-35B-A3B (same)LoRA SFT with text-only Claude reasoning chains

Source

This model was converted to MLX format from FINAL-Bench/Darwin-35B-A3B-Opus using mlx-vlm version 0.4.4.

Visit Website

0 reviews

5
0
4
0
3
0
2
0
1
0
Likes11
Downloads
📝

No reviews yet

Be the first to review TheCluster/Darwin-35B-A3B-Opus-MLX-mxfp8!

Model Info

ProviderTheCluster
Categoryimage
Reviews0
Avg. Rating / 5.0

Community

Likes11
Downloads

Rating Guidelines

★★★★★Exceptional
★★★★Great
★★★Good
★★Fair
Poor