TinmanLabSL commited on
Commit
eef6293
·
verified ·
1 Parent(s): cd599b6

Refresh org card: DD paper published, list public PoC models post-migration

Browse files
Files changed (1) hide show
  1. README.md +63 -41
README.md CHANGED
@@ -1,41 +1,63 @@
1
- ---
2
- title: README
3
- emoji: 🔩
4
- colorFrom: gray
5
- colorTo: blue
6
- sdk: static
7
- pinned: false
8
- ---
9
-
10
- <div align="center">
11
-
12
- # Tinman Lab
13
-
14
- ### Autonomous Machines. Second-Order Systems.
15
-
16
- <sub>AGENT MEMORY · ADVERSARIAL SAFETY · AGENTIC ECONOMY · PERCEPTION SYSTEMS · APPLIED RESEARCH</sub>
17
-
18
- ---
19
-
20
- </div>
21
-
22
- ## Disposition Distillation
23
-
24
- Tinman Lab develops **Disposition Distillation (DD)** — a multi-teacher distillation methodology that trains *how a model behaves* into weights, not system prompts. DD models plan before acting, acknowledge uncertainty, verify their own reasoning, and know what they don't know.
25
-
26
- - **4-stage all-MIT pipeline** — Kimi K2.5 → GLM-5 → MiniMax M2.7 → GLM-5
27
- - **7 behavioral dispositions** Eager, Deliberate, Adversarial, Curious, Self-Improving, Humble, Persistent
28
- - **On-device focus** — 0.6B to 2B parameters, quantized for mobile and edge deployment
29
- - **100% open training data** MIT-licensed teachers only, zero proprietary model outputs
30
-
31
- ## Models
32
-
33
- | Model | Size | Description |
34
- |-------|------|-------------|
35
- | [tinman-code-0.6B](https://huggingface.co/Tinman-Lab/tinman-code-0.6B) | 418 MB | Coding assistant with meta-cognitive awareness plans, verifies, flags uncertainty |
36
-
37
- ## Links
38
-
39
- - [Website](https://tinmanlab.com)
40
- - [GitHub](https://github.com/tinmanlabsl/)
41
- - Research Paper — coming soon
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ title: README
3
+ emoji: 🔩
4
+ colorFrom: gray
5
+ colorTo: blue
6
+ sdk: static
7
+ pinned: false
8
+ ---
9
+
10
+ <div align="center">
11
+
12
+ # Tinman Lab
13
+
14
+ ### Autonomous Machines. Second-Order Systems.
15
+
16
+ <sub>AGENT MEMORY · ADVERSARIAL SAFETY · AGENTIC ECONOMY · PERCEPTION SYSTEMS · APPLIED RESEARCH</sub>
17
+
18
+ ---
19
+
20
+ </div>
21
+
22
+ ## Research
23
+
24
+ ### Disposition Distillation — a three-arc negative result
25
+
26
+ We set out to train *behavioral dispositions* — self-verification, uncertainty
27
+ acknowledgment, feedback integrationinto sub-billion-parameter language
28
+ models. Across three independent operator classes (SFT/DPO LoRA imitation,
29
+ attention-head tempering on `o_proj`, and frozen-base hidden-state confidence
30
+ sidecars), no operator moved judge-measured disposition without simultaneously
31
+ damaging content quality or collapsing into stylistic mimicry. The failure is
32
+ consistent across Qwen3-0.6B, Qwen3-1.7B, Qwen3.5-0.8B, Gemma 4 E2B, and
33
+ SmolLM2-1.7B-Instruct.
34
+
35
+ The contribution is the falsification, a two-failure-mode taxonomy for linear
36
+ hidden-state probes, and a methodological pipeline that converts
37
+ CV-on-same-distribution false positives into honest negatives.
38
+
39
+ 📄 **Paper** — [arXiv:2604.11867](https://arxiv.org/abs/2604.11867)
40
+ 🔧 **Artifacts** — [github.com/tinmanlabsl/disposition-distillation](https://github.com/tinmanlabsl/disposition-distillation)
41
+
42
+ ## Open-source proofs of concept
43
+
44
+ Apache 2.0. Internal product models are not listed here.
45
+
46
+ ### Tinman SmolOmni (MLA)
47
+
48
+ - [Tinman-SmolOmni-MLA-256M](https://huggingface.co/Tinman-Lab/Tinman-SmolOmni-MLA-256M)
49
+ - [Tinman-SmolOmni-MLA-500M](https://huggingface.co/Tinman-Lab/Tinman-SmolOmni-MLA-500M)
50
+ - [Tinman-SmolOmni-MLA-Toolkit](https://huggingface.co/Tinman-Lab/Tinman-SmolOmni-MLA-Toolkit)
51
+
52
+ ### Tinman Companion (Gemma 4)
53
+
54
+ - [Tinman-gemma4-companion-merged](https://huggingface.co/Tinman-Lab/Tinman-gemma4-companion-merged) — full precision
55
+ - [Tinman-gemma4-companion-gguf](https://huggingface.co/Tinman-Lab/Tinman-gemma4-companion-gguf) — GGUF quantized for llama.cpp
56
+ - [Tinman-gemma4-companion-litert-lm](https://huggingface.co/Tinman-Lab/Tinman-gemma4-companion-litert-lm) — LiteRT-LM for on-device deployment
57
+ - [Tinman-gemma4-companion-sft](https://huggingface.co/Tinman-Lab/Tinman-gemma4-companion-sft) — SFT checkpoint
58
+ - [Tinman-gemma4-companion-dpo](https://huggingface.co/Tinman-Lab/Tinman-gemma4-companion-dpo) — DPO checkpoint
59
+
60
+ ## Links
61
+
62
+ - [Website](https://tinmanlab.com)
63
+ - [GitHub](https://github.com/tinmanlabsl/)