--- license: apache-2.0 base_model: Qwen/Qwen3-1.7B tags: - qwen3 - fine-tuned - hito - hitonet - reasoning - conversational - thinking - adaptive-reasoning - tree-of-thought - hierarchical-reasoning - cognitive-framework - self-aware-ai - anti-hallucination - synthetic-data - gguf - llama-cpp - ollama pipeline_tag: text-generation language: - en library_name: transformers ---
Hitonet Meet Hito # Hito 1.7B ### Brain, Heart, and a Really Good Memory [![Website](https://img.shields.io/badge/hitonet.com-000000?style=for-the-badge&logo=globe&logoColor=white)](https://hitonet.com) [![Chat](https://img.shields.io/badge/Try_Free_Chat-22c55e?style=for-the-badge&logo=chatbot&logoColor=white)](https://chat.hitonet.com) [![API](https://img.shields.io/badge/API_Platform-3b82f6?style=for-the-badge&logo=swagger&logoColor=white)](https://platform.hitonet.com) [![Pricing](https://img.shields.io/badge/Pricing-8b5cf6?style=for-the-badge&logo=stripe&logoColor=white)](https://platform.hitonet.com/pricing) --- Status Parameters Training Context License
--- > [!NOTE] > **EXPERIMENTAL MODEL - PROOF OF CONCEPT** > > This 1.7B model was fine-tuned on just **~300 examples** generated by **Hito-Genius** (our flagship model). It's an experiment in knowledge distillation - can a tiny model learn to think like a bigger one? > > **Don't expect production quality.** This is proof that the cognitive architecture transfers, not a production release. > > For the real deal, use our API at [platform.hitonet.com](https://platform.hitonet.com). --- ## ๐Ÿงช The Experiment **Question:** Can we teach a 1.7B model to think like our flagship Hito-Genius? **Method:** Generate ~300 high-quality reasoning examples from Hito-Genius, fine-tune a small model on them. **Result:** It actually works. Kind of. The cognitive patterns transfer, even with minimal data. | What This Proves | What This Doesn't Prove | |------------------|-------------------------| | Cognitive architecture can be distilled | That 300 examples is enough | | Small models can learn structured thinking | That this is production-ready | | Tree-reasoning transfers from teacher | That it matches Hito-Genius quality | --- ## ๐Ÿ“ˆ Benchmark Results (December 2025) We tested Hito 1.7B against leading small models on counting, math, and self-awareness tasks.
Size vs Performance
### Summary Results | Model | Params | Accuracy | Counting | Math | |-------|--------|----------|----------|------| | GPT-5-mini | ~8B | **100%** | 100% | 100% | | Claude Haiku 4.5 | ~8B | 90% | 67% | 100% | | **Hito 1.7B** | **1.7B** | **80%** | **67%** | **100%** | | GPT-4o-mini | ~8B | 80% | 33% | 100% | | Claude 3.5 Haiku | ~8B | 70% | 33% | 100% | | Qwen3 1.7B base | 1.7B | 17% | 0% | 17% | ### The Famous Strawberry Test *"How many r's are in 'strawberry'?"*
Counting Comparison
| Model | Answer | Correct | |-------|--------|---------| | **Hito 1.7B** | **3** | โœ… | | Qwen3 1.7B (base) | 2 | โŒ | | GPT-4o-mini | 2 | โŒ | | Claude 3.5 Haiku | 2 | โŒ | **Hito 1.7B solved the counting problem that larger models failed!** ### Why? The `` Tag in Action ```xml Let me spell it out: s-t-r-a-w-b-e-r-r-y Counting r's: position 3 (r), position 8 (r), position 9 (r) Total: 3 3 ``` The cognitive training teaches the model to **verify** instead of guessing. --- ## ๐Ÿ“š Prior Work (Being Honest) We didn't invent thinking in AI. Here's what came before us: | Research | What They Did | How Hito Differs | |----------|---------------|------------------| | **Chain-of-Thought** (Wei et al., 2022) | Prompting with "Let's think step by step" | We TRAIN the model to think, not just prompt | | **OpenAI o1/o3** (2024-2025) | Hidden thinking tokens | Our thinking is TRANSPARENT and OPEN | | **Reflexion** (Shinn et al., 2023) | Agents reflecting on mistakes | Self-reflection is IN the weights, not external | | **Tree of Thoughts** (Yao et al., 2023) | Branching paths via search | Our branching is LEARNED, not algorithmic | | **Emotional AI** (WASABI, BELBIC) | Emotion classification/simulation | We simulate emotional CONTEXT in responses | ### What Makes Hito Different? 1. **Combined Approach**: Cognitive + emotional + self-doubt in ONE framework 2. **Tiny Model**: 1.7B params, not 100B+ 3. **Open Weights**: Run locally, see how it thinks 4. **Trained, Not Prompted**: Behavior is in the weights 5. **Humble by Design**: Says "I might be wrong" when uncertain *We stand on the shoulders of giants. Our contribution is making these techniques accessible in a small, open model.* --- ## ๐Ÿ“Š Training Details | Property | Value | |----------|-------| | **Base Model** | Qwen/Qwen3-1.7B | | **Training Examples** | ~300 | | **Data Source** | Generated by Hito-Genius | | **Method** | Supervised Fine-Tuning (SFT) | | **Purpose** | Proof of Concept | *Yes, only 300 examples. We wanted to see how far we could push minimal data with high-quality synthetic examples.* --- ## ๐ŸŽฏ The Problem We're Solving Most AI models are **confidently wrong**. They hallucinate, make up facts, and never question themselves. **We're fixing this by teaching AI to understand its own limitations.** --- ## ๐Ÿ” Hito Knows Its Weaknesses | Limitation | Why It Happens | How Hito Handles It | |------------|----------------|---------------------| | **Can't count reliably** | "I process tokens, not characters." | Numbers each item, counts backwards to verify | | **Math errors** | "I don't have a calculator." | Writes out every step instead of mental math | | **Hallucination** | "I can make up false information." | Uses `` and `` tags | | **Overconfidence** | "I can sound sure when wrong." | `` tag rates certainty | ### Example: Self-Correcting Math ```xml 15% of 200 = 15 ร— 200 = 3000 Wait... that's way too high for a percentage. I multiplied instead of calculating percentage. 15% = 0.15 0.15 ร— 200 = 30 โœ“ ``` --- ## ๐Ÿง  Cognitive Architecture *Distilled from Hito-Genius into this tiny model.*
Cognitive Architecture
### Four Cognitive States | State | Focus | |-------|-------| | **Analytical** | Logic, accuracy | | **Creative** | Imagination, exploration | | **Empathetic** | Feelings, perspectives | | **Reflective** | Depth, meaning | --- ## ๐ŸŒณ Tree-Structured Reasoning Not linear chain-of-thought. Tags **nest**, **branch**, and **recurse**.
Tree-Structured Reasoning
--- ## ๐ŸŽจ Creative Flow
Creative Flow
--- ## ๐Ÿ›ก๏ธ The Humble Tags | Tag | Purpose | |-----|---------| | `` | Question assumptions | | `` | Admit errors | | `` | Acknowledge gaps | | `` | Rate certainty | | `` | Double-check work | --- ## ๐Ÿ“ฆ Available Files | File | Size | |------|------| | `hito-1.7b-q8_0.gguf` | **1.8 GB** (recommended) | | `hito-1.7b-f16.gguf` | 3.3 GB | | `model.safetensors` | 3.4 GB | --- ## โšก Quick Start ### Ollama ```bash wget https://huggingface.co/hitonet/hito-1.7b/resolve/main/hito-1.7b-q8_0.gguf cat > Modelfile << 'EOF' FROM hito-1.7b-q8_0.gguf SYSTEM "You are Hito by Hitonet.com." PARAMETER temperature 0.7 PARAMETER stop "<|im_end|>" EOF ollama create hito -f Modelfile ollama run hito ``` ### API (The Real Hito-Genius) ```bash curl https://hitonet.com/v1/chat/completions \ -H "Authorization: Bearer YOUR_API_KEY" \ -H "Content-Type: application/json" \ -d '{"model": "hito-genius", "messages": [{"role": "user", "content": "Hello!"}]}' ``` Try the real thing at [platform.hitonet.com](https://platform.hitonet.com) โ€” $1 free credit! --- ## ๐Ÿ”ฎ What's Coming This 1.7B experiment proves the concept. Our **foundational model** is in development: - Full cognitive architecture at scale - Thousands of training examples - Production-ready reliability - The next evolution of Hito *This is just the beginning.* ---
**Made with genuine curiosity by [Hitonet](https://hitonet.com)** *Trained on 300 examples. Learned to doubt itself. That's pretty cool.*