π§ ZeroXClem/Qwen3-1.7B-TardigradePro
A resilient and highly distilled 1.7B parameter model crafted through precision mergekit stock fusion, combining powerful reasoning, advanced code generation, and symbolic logic.
π§ Merge Configuration
name: ZeroXClem/Qwen3-1.7B-TardigradePro
base_model: Qwen/Qwen3-1.7B-Base
dtype: bfloat16
merge_method: model_stock
models:
- model: prithivMLmods/Capricornus-MoT-1.7B-Supreme1
- model: ertghiu256/qwen3-1.7b-mixture-of-thought
- model: XformAI-india/qwen-1.7b-coder
- model: prithivMLmods/Regulus-Qwen3-R1-Llama-Distill-1.7B
- model: prithivMLmods/Demeter-LongCoT-Qwen3-1.7B
tokenizer_source: Qwen/Qwen3-1.7B-Base
π Key Strengths
𧬠Multi-Expert Fusion Integrates chain-of-thought, symbolic reasoning, long-context STEM tasks, and multilingual coding into a unified agent.
π‘ Lightweight Distilled Intelligence Powered by distilled traces and Mixture-of-Thoughts tuning, delivering performance far beyond its 1.7B parameter size.
π Domain-Specific Superpowers Excels in:
- Math & science step-by-step reasoning
- Python/JS/Bash code generation
- Chain-of-thought problem decomposition
- Structured output: LaTeX, JSON, Markdown, YAML
π§ LongCoT Capabilities Handles deep reasoning flows with clarityβideal for algorithm design, math derivations, and structured academic workflows.
β‘ Efficiency at Scale Deployable on 8GB VRAM or CPU edge devices with GGUF and quantized variants.
π Model Lineage
| Model | Contribution |
|---|---|
prithivMLmods/Capricornus-MoT-1.7B-Supreme1 |
MoT fine-tuning on code/math/science |
ertghiu256/qwen3-1.7b-mixture-of-thought |
Multiexpert symbolic logic |
XformAI-india/qwen-1.7b-coder |
Code specialization (Python, JS, Bash) |
prithivMLmods/Regulus-Qwen3-R1-Llama-Distill-1.7B |
DeepSeek70B reasoning distilled |
prithivMLmods/Demeter-LongCoT-Qwen3-1.7B |
Long chain-of-thought symbolic tuning |
Qwen/Qwen3-1.7B-Base |
Foundational pretraining and multilingual context |
π§ͺ Quickstart
from transformers import AutoTokenizer, AutoModelForCausalLM
model = AutoModelForCausalLM.from_pretrained("ZeroXClem/Qwen3-1.7B-TardigradePro", device_map="auto", torch_dtype="auto")
tokenizer = AutoTokenizer.from_pretrained("ZeroXClem/Qwen3-1.7B-TardigradePro")
prompt = "Explain with code how to solve a quadratic equation and show symbolic math."
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
π» Applications
- Educational assistants and STEM tutors
- Coding copilots and REPL automation
- Symbolic math & logic agents
- Scientific document assistants
- Edge-deployable AI reasoning models
π« Limitations
- Context limited to ~8K - 30K Tokens
- Specialized for structured logicβless effective in freeform chat
- Doesnβt hallucinate creativelyβfocuses on truth, clarity, and structure
π Related Resources
π License
Apache 2.0 Portions inherit wen licensesβsee each base model for additional usage terms.
π Credits
ZeroXClem Model Made with π Special Thanks to prithivMLmods, ertghiu256, XformAI, and Qwen Merging, distillation, prompt tuning and symbolic alignment by love and fire.
βBuilt like a tardigradeβsmall, mighty, and immortal.β
- Downloads last month
- 18
