README / README.md
sip786's picture
align HF org card with API private-beta truth (post-Natalie API report)
3b6ced5 verified
metadata
title: Sipsa Labs
emoji: 🧊
colorFrom: indigo
colorTo: blue
sdk: static
pinned: false

Sipsa Labs, Inc.

Sipsa Labs, Inc. is an experimental and deep tech-and-software company (Delaware C-corp, incorporated May 2026). We invent and ship across the full breadth of tech and software β€” deep research, runtime systems, novel substrates, infrastructure, hardware-adjacent stacks, and software products that don't fit anywhere else yet. UltraCompress β€” lossless 5-bit transformer compression β€” is our first flagship publicly-shipped product. More products in flight.


2026-05-11 β€” 22 architectures verified at 5 bpw, three new sub-1.005Γ— records this week

Live on Hacker News today: https://news.ycombinator.com/item?id=48099107

Production-grade lossless 5-bit transformer compression across 22 architectures, dense + Mixture-of-Experts + state-space (Mamba), 0.6B to 405B parameters. Mathematically lossless customer-side reconstruction: SHA-256 over reconstructed tensor bytes matches the trainer'''s measurement, verified by uc verify.

Tightest verified PPL ratios at 5 bpw

(perplexity ratio = compressed PPL / bf16 baseline PPL; FineWeb-edu held-out tail; seq_len = 1024; n = 30-50; seed = 42)

Model Params Type PPL ratio Notes
Phi-3-mini-4k-instruct 3.8B dense 1.00262Γ— seq_len=128 caveat
Mixtral-8x7B 47B MoE 1.00368Γ— tightest MoE result
Qwen3-1.7B-Base 1.7B dense 1.00401Γ— small-decoder record
Qwen3-14B 14B dense 1.00403Γ— 14B-class record
Yi-1.5-9B 8.8B dense 1.00414Γ— >8B record
Qwen3-8B 8B dense 1.00440Γ— 8B-class record
Mistral-7B-v0.3 7B dense 1.00548Γ— NEW this week β€” 9.16Γ— tighter than prior
Phi-3-mini-4k 3.8B dense 1.00624Γ— cross-arch confirm at seq=1024
Hermes-3-Llama-3.1-405B 405B dense 1.0066Γ— largest dense 5-bit lossless on the Hub
Qwen3-0.6B 0.6B dense 1.0069Γ—
OLMo-2-0425-1B 1B dense 1.0073Γ—
SmolLM2-1.7B-Instruct 1.7B dense 1.0075Γ—
SmolLM2-1.7B 1.7B dense 1.0085Γ—
Mamba-2.8B 2.8B SSM 1.0119Γ— first published 5-bit lossless on a state-space model
Llama-3.1-8B 8B dense 1.0125Γ— standard eval

Quick start (3 commands, no GPU required for verify)

pip install ultracompress
hf download SipsaLabs/qwen3-8b-uc-v3-bpw5 --local-dir ./qwen3-8b
uc bench ./qwen3-8b

Inference API

api.sipsalabs.com API is in private beta, OpenAI-compatible:

export OPENAI_BASE_URL=https://api.sipsalabs.com/v1
# The official `openai` SDK works unchanged.

Pricing tiers + Compression-as-a-Service contracts at https://sipsalabs.com/pricing.

License + IP

  • PyPI ultracompress v0.6+ under BUSL-1.1 with Additional Use Grant β€” free for sub-$1M ARR companies + research + individuals. Auto-converts to Apache 2.0 four years after each release.
  • v0.5.x stays Apache-2.0 forever on legacy/0.5.x.
  • Two USPTO provisionals filed 2026-04-25: 64/049,511 + 64/049,517.

Commercial inquiries: founder@sipsalabs.com

Public verifier dashboard: https://sipsalabs.com/inference
Selective Disclosure Charter β€” what we publish vs what we keep internal: see github.com/sipsalabs/ultracompress