DistilQwen Collection Two students, one methodology. 30B teacher → 1.7B and 0.6B via proof-weighted distillation + legal SFT. Six models, Apache 2.0. • 9 items • Updated about 4 hours ago
DistilQwen Collection Two students, one methodology. 30B teacher → 1.7B and 0.6B via proof-weighted distillation + legal SFT. Six models, Apache 2.0. • 9 items • Updated about 4 hours ago
reaperdoesntknow/Qwen3-1.7B-Coder-Distilled-SFT-GGUF Text Generation • 2B • Updated about 20 hours ago • 22
reaperdoesntknow/Qwen3-1.7B-Coder-Distilled-SFT-GGUF Text Generation • 2B • Updated about 20 hours ago • 22
reaperdoesntknow/Qwen3-1.7B-Coder-Distilled-SFT Text Generation • 2B • Updated about 20 hours ago • 58
reaperdoesntknow/Qwen3-1.7B-Coder-Distilled-SFT Text Generation • 2B • Updated about 20 hours ago • 58
DistilQwen Collection Two students, one methodology. 30B teacher → 1.7B and 0.6B via proof-weighted distillation + legal SFT. Six models, Apache 2.0. • 9 items • Updated about 4 hours ago
reaperdoesntknow/Qwen3-0.6B-Distilled-30B-A3B-Thinking-SFT-GGUF Text Generation • 0.8B • Updated 4 days ago • 190
reaperdoesntknow/Qwen3-0.6B-Distilled-30B-A3B-Thinking-SFT-GGUF Text Generation • 0.8B • Updated 4 days ago • 190
reaperdoesntknow/Qwen3-0.6B-Distilled-30B-A3B-Thinking-SFT Text Generation • 0.8B • Updated 4 days ago • 32 • 1
reaperdoesntknow/Qwen3-0.6B-Distilled-30B-A3B-Thinking-SFT Text Generation • 0.8B • Updated 4 days ago • 32 • 1