KyleHessling1/Qwopus-GLM-18B-Merged-GGUF
Summary
An experimental 18B-parameter model created by stacking two Qwen-3.5-9B finetunes and healing the layer boundary with 1000-step QLoRA; the resulting GGUF beats Qwen 3.6-35B MoE on a 44-test suite while fitting in 9.2 GB VRAM.
View Cached Full Text
Cached at: 04/21/26, 01:37 PM
KyleHessling1/Qwopus-GLM-18B-Merged-GGUF · Hugging Face
Source: https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#qwopus-glm-18b-merged-healedQwopus-GLM-18B-Merged (Healed)
A64-layer frankenmergeof two ofJackrong’sincredible Qwen3.5-9B finetunes, stacking all 32 layers from each to create an ~18B parameter model, thenhealed with a 1000-step QLoRA fine-tuneto smooth the layer boundary.
**This was a fun experiment!**A lot of people have been asking for something between Jackrong’s 27B and 9B models — something that runs well on 12–16 GB GPUs. This frankenmerge is an attempt at filling that gap, and the results are surprisingly good.
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#heal-fine-tune–it-worksHeal Fine-Tune — It Works
The raw frankenmerge had a known issue:garbled code output. Because two separately-trained models were stacked at layer 32, structured output (code blocks, HTML, bracket matching) would occasionally come out malformed or hallucinated.
We ran a1000-step QLoRA heal fine-tuneusing Jackrong’s own training data to let gradients flow across the layer boundary — and the results are significant:
- HTML generation is now clean and production-quality.We tested a complex single-page weather dashboard (navbar, dark mode toggle, 5-day forecast grid, responsive sidebar, CSS variables, JavaScript) — the model produced 14,500+ chars of valid HTML/CSS/JS withperfectly balanced CSS braces, perfectly balanced JS parentheses, no garbled text, and a complete
</html\>closure. - **Programming benchmark improved:**11/15 (raw) -> 12/15 (healed), recovering the
longest\_substringsliding-window algorithm test (8/8 cases passing) - **Overall score improved:**39/44 ->40/44 (90.9%), still comfortably beating Qwen 3.6 MoE (38/44) at less than half the VRAM
- **Loss dropped 39%**during training (1.02 -> 0.62), confirming the layer boundary was a real source of error that training could address
The healed GGUF (Qwopus\-GLM\-18B\-Healed\-Q4\_K\_M\.gguf) is the only version in this repo. If you’re interested in the raw unhealed merge for research purposes, reach out.
This is still an experimental model — it may have quirks or issues. If you run into anything weird, or if you make something cool with it, reach out on X:@KyleHessling1
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#benchmark-resultsBenchmark Results
We ran a 44-test capability suite covering basic generation, reasoning, tool calling, agentic workflows, structured output, context handling, multilingual, programming, and performance.
The healed mergeoutperforms the brand new Qwen 3.6-35B-A3B MoE(Q4_K_M, 22 GB) despite being significantly smaller (Q4_K_M, 9.2 GB):
CategoryQwopus 9B (source)Qwopus-GLM-18B (healed)Qwen 3.6-35B MoEBasic6/66/65/6Reasoning4/44/44/4Tool Calling6/66/66/6Agentic4/44/44/4Structured Output2/22/22/2Context2/32/32/3Multilingual2/22/22/2Programming13/1512/1512/15Performance2/22/21/2TOTAL****41/44 (93.2%)****40/44 (90.9%)****38/44 (86.4%)Throughput126.0 tok/s66.0 tok/s174.2 tok/sGGUF Size5.3 GB9.2 GB22 GB
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#key-takeawaysKey Takeaways
- 40/44 tests passed (90.9% healed)— beats Qwen 3.6 MoE’s 38/44 (86.4%) at less than half the VRAM
- Heal training recovered programming capability: 11/15 raw -> 12/15 healed (matching Qwen 3.6 MoE)
- Perfect tool calling (6/6)— single calls, optional params, tool selection, complex params, response handling
- Perfect agentic reasoning (4/4)— plan generation, multi-step tool workflows, error recovery, self-correction
- Highest Chinese output densityof any model tested: 129-138 CJK characters
- ~66 tok/swith low throughput variance — stable inference
- Fits in 12 GB VRAMat Q4_K_M — runs on consumer GPUs like RTX 3060/4070
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#heal-fine-tune-detailsHeal Fine-Tune Details
The raw frankenmerge had code formatting issues (garbled code blocks, missing brackets). We ran a 1000-step QLoRA heal fine-tune using Jackrong’s training data to smooth the layer-32 boundary:
- **Method:**QLoRA (4-bit NF4), LoRA rank 64, targeting all attention + MLP projections
- **Data:**Blend of
Jackrong/Qwen3\.5\-reasoning\-700x(70%),Jackrong/Competitive\-Programming\-python\-blend(15%),Jackrong/MultiReason\-ChatAlpaca(15%) - **Training:**1000 steps, batch 8, lr 2e-5 cosine, ~14 hours on RTX 5090
- **Loss:**1.02 -> 0.62 (39% reduction)
- **Result:**Recovered 1 programming test, HTML/CSS output is now clean and production-quality
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#where-it-falls-shortWhere It Falls Short
Three programming tests still fail on the healed version: one function naming issue, one missing JS paren, and one that doesn’t produce a code block for pytest generation. These are residual formatting artifacts from the merge.
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#frontend-code-generation–stress-test-resultsFrontend Code Generation — Stress Test Results
We put the healed model through a rigorous frontend stress test: 6 increasingly complex HTML/CSS/JS generation tasks, each requiring thousands of tokens of structurally valid code output. The results speak for themselves:
TestWhat We Asked ForChecks PassedOutput SizeWeather DashboardResponsive dashboard, CSS vars, dark mode toggle, 5-day forecast grid9/914.5K charsE-Commerce Product PageImage gallery, color swatches, quantity selector, tabbed content, sticky mobile bar12/1216.7K charsAnimated SaaS LandingMoving gradient, typing animation, IntersectionObserver scroll reveals, auto-rotating testimonial carousel, 3 pricing tiers13/1324.1K charsAnalytics DashboardSVG bar chart with tooltips, SVG donut chart, sortable data table, collapsible sidebar, dark theme13/1322.3K charsMulti-Step Registration3-step form wizard, real-time validation, password strength meter, state dropdown, animated transitions, success modal12/1223.3K charsSnake GameCanvas game loop, arrow key controls, collision detection, localStorage high score, increasing difficulty11/1211.2K chars 62/63 total checks passed (98.4%)
Every single output had:
- Perfectly balanced CSS braces(zero imbalance across all 6 files)
- Perfectly balanced JS parentheses(zero imbalance across all 6 files)
- Zero garbled or hallucinated text
- Working JavaScript— dark mode toggles, IntersectionObserver animations, SVG chart rendering, form validation, canvas game loops
The only miss: the Snake game had a minor closing tag typo (html\>instead of</html\>) at the very end.
This is remarkable for a frankenmerge of two 9B models with only 1000 steps of QLoRA healing. The model is producingproduction-quality frontend code— not just syntactically valid HTML, but sophisticated interactive applications with modern CSS (Grid, Flexbox, custom properties, keyframe animations) and non-trivial JavaScript (IntersectionObserver, requestAnimationFrame game loops, real-time form validation, SVG chart generation).
All 6 sample HTML files are included in thesamples/directory of this repo — download them and open in a browser to see for yourself.
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#architectureArchitecture
PropertyValueTotal Layers64 (32 + 32)Total Parameters~18BHidden Size4096Attention Heads16 (4 KV heads, GQA)Intermediate Size12288Context Length262,144 tokensAttention TypeHybrid (linear + full, every 4th layer)GGUF Q4_K_M Size9.2 GB
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#layer-compositionLayer Composition
Layers 0–31: Jackrong/Qwopus3.5-9B-v3.5 (Opus reasoning distill)
Layers 32–63: Jackrong/Qwen3.5-9B-GLM5.1-Distill-v1 (GLM-5.1 reasoning distill)
Embeddings, LM head, MTP, vision encoder: from Qwopus3.5-9B-v3.5
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#source-modelsSource Models
All credit for the source models goes to**Jackrong**, who created both of these excellent finetunes. I just stacked the layers — the quality comes from his work.
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#jackrongqwopus35-9b-v35Jackrong/Qwopus3.5-9B-v3.5
A reasoning-enhanced finetune of Qwen3.5-9B trained with ~2x more SFT data than v3, focused on structured reasoning, tool-augmented workflows, and multi-step agentic tasks.
Key insight from the v3.5 design:*“Scaling high-quality SFT data may further enhance the generalization ability of large language models.”*Reasoning SFT helps models better utilize existing knowledge and activate latent knowledge through structured reasoning, rather than simply memorizing long Chain-of-Thought outputs.
Performance highlights (27B line reference):
- MMLU-Pro: 90.36% accuracy (+1.07% over v3)
- Agentic coding tests: 43/44 passed (97.7%)
Resources:
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#jackrongqwen35-9b-glm51-distill-v1Jackrong/Qwen3.5-9B-GLM5.1-Distill-v1
A distilled variant of Qwen3.5-9B trained on high-quality reasoning data from a GLM-5.1 teacher model (~700x scale of Qwen3.5-reasoning-700x). Focused on structured reasoning ability, instruction-following consistency, and problem decomposition.
The model learns a structured reasoning scaffold: understand the task, break down the problem, reason step by step, then construct the final answer.
Training data:
- Primary:
Jackrong/GLM\-5\.1\-Reasoning\-1M\-Cleaned(cleaned from Kassadin88/GLM-5.1-1000000x) - Auxiliary:
Jackrong/Qwen3\.5\-reasoning\-700x
Resources:
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#why-this-worksWhy This Works
By stacking two differently-distilled reasoning models, this merge combines:
- Qwopus v3.5’s strengthsin agentic tool use, code generation, and token-efficient reasoning (Opus-style training)
- GLM-5.1 Distill’s strengthsin structured problem decomposition, instruction adherence, and chain-of-thought organization (GLM-style reasoning scaffold)
The hypothesis: deeper networks with diverse reasoning training produce more robust, capable models — and the benchmark results suggest it works, at least for the capabilities we tested.
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#merge-detailsMerge Details
- **Method:**Passthrough frankenmerge (layer stacking)
- **Tool:**Custom script (mergekit did not support Qwen3.5’s hybrid linear/full attention architecture)
- **Embeddings / LM Head / Visual / MTP:**From Qwopus3.5-9B-v3.5
- **Precision:**BF16 -> Q4_K_M GGUF
- No additional trainingwas performed
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#usageUsage
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#with-llamacpp-recommendedWith llama.cpp (recommended)
llama-server \
-m Qwopus-GLM-18B-Merged-Q4_K_M.gguf \
--chat-template-file your-qwen35-template.jinja \
--ctx-size 65536 \
--flash-attn on \
--n-gpu-layers 99
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#with-transformersWith Transformers
The full BF16 safetensors are not included in this GGUF repo. If you need them for further fine-tuning or experimentation, reach out on X.
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#limitationsLimitations
- Experimental frankenmerge— no additional training was done. Layer boundaries between the two source models may cause occasional coherence or formatting issues.
- Code formatting— the model sometimes garbles fenced code blocks (returns code without proper markdown fencing). The reasoning is usually correct even when the formatting isn’t.
- Not exhaustively tested— this was a fun weekend project. There may be edge cases we haven’t found yet.
- Hallucination risk— as with all autoregressive LLMs, outputs may contain factual errors.
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#supported-researchSupported Research
Both source models reference:Ren et al., 2026 —Rethinking Generalization in Reasoning SFT(arXiv:2604.06628)
Key findings: reasoning SFT generalizes when sufficiently trained; high-quality long-CoT data enables cross-domain transfer; stronger models learn reasoning structure, not just longer outputs.
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#acknowledgementsAcknowledgements
- Jackrong— the real MVP. Both source models, training pipelines, datasets, and documentation are his work. This merge exists because his finetunes are so good that even stacking them naively produces something surprisingly capable.
- **Qwen**for the excellent Qwen3.5-9B base model
- **Unsloth AI**for efficient fine-tuning infrastructure
- GLM-5.1 teamfor the teacher model used in distillation
- **Kassadin88**for the original GLM-5.1-1000000x dataset
- The broader open-source community
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#questionsQuestions?
This was just for fun — reach out on X if you have questions, find issues, or build something cool with it!
https://huggingface.co/KyleHessling1/Qwopus-GLM-18B-Merged-GGUF#citationsCitations
@misc{jackrong_qwopus35_9b_v35,
title = {Qwopus3.5-9B-v3.5},
author = {Jackrong},
year = {2026},
publisher = {Hugging Face}
}
@misc{jackrong_qwen35_9b_glm51_distill_v1,
title = {Qwen3.5-9B-GLM5.1-Distill-v1},
author = {Jackrong},
year = {2026},
publisher = {Hugging Face}
}
Similar Articles
Jackrong/Qwopus-GLM-18B-Merged-GGUF
Jackrong released Qwopus-GLM-18B-Merged-GGUF, a 64-layer frankenmerge combining two Qwen3.5-9B finetunes into an ~18B parameter model, healed with 1000-step LoRA fine-tuning to fix layer boundary issues. The model achieves 90.9% on capability benchmarks while using less than half the VRAM of Qwen 3.6-35B MoE.
@outsource_: NEW GLM+ QWEN 18B RUNS ON CONSUMER GPU IT BEATS 35B MoE AT HALF THE VRAM @KyleHessling1 just dropped the healed Qwopus-…
A new 18B merged quantized model, Qwopus-GLM-18B-GGUF, outperforms 35B MoE models while using half the VRAM and running on consumer GPUs.
hesamation/Qwen3.6-35B-A3B-Claude-4.6-Opus-Reasoning-Distilled-GGUF
A 35B-parameter Qwen3.6 model fine-tuned with Claude-Opus-style chain-of-thought distillation data and released in GGUF quantized formats for efficient local inference.
DavidAU/Qwen3.6-27B-Heretic-Uncensored-FINETUNE-NEO-CODE-Di-IMatrix-MAX-GGUF
A community-finetuned, uncensored version of the Qwen 3.6 27B model featuring high-precision GGUF quantizations.
Qwen3.6-27B-GGUF is here!
Community GGUF release of Qwen’s 27B hybrid-architecture model with 262k context, multimodal inputs, tool calling and "Thinking Preservation" for agentic coding.