Tear Labs

Live now Public release 2.5-bit kernels vLLM-ready

AI so cheap, you'll cry.

State-of-the-art inference. Embarrassingly affordable.

01 Live now

Qwen3.5 0.8B, 4B, and 9B at 2.5-bit

Public release note

Perplexity holds within 2% of original and the drop ships with kernels for vLLM.

0.8B / 4B / 9B2.5-bit1.9 GB on 4B
Read release note ↗
02 Queued

Second public release

The next drop is already staged while validation finishes across realistic batch sizes, packaging, and serving checks.

Watch the blog ↗

Arrived.

New drops, straight to your inbox.

No spam. One click to unsubscribe.