Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up
robtacconelliΒ 
posted an update 2 days ago
Post
3599
🧬 Midicoth: diffusion-based lossless compression β€” no neural net, no GPU, no training data

What if reverse diffusion could compress text β€” without a neural network?
Midicoth brings score-based denoising into classical compression. It treats prior smoothing as forward noise and reverses it with Tweedie's formula on a binary tree β€” 3 denoising steps, James-Stein shrinkage, applied after all model blending. ~2,000 lines of C, single CPU core.

Beats every dictionary compressor we tested:
enwik8 (100 MB) β†’ 1.753 bpb (βˆ’11.9% vs xz, βˆ’15% vs Brotli, βˆ’24.5% vs bzip2)
alice29.txt β†’ 2.119 bpb (βˆ’16.9% vs xz)
Outperforms xz, zstd, Brotli, bzip2, gzip on all inputs

PAQ/CMIX still win with hundreds of models + LSTMs. LLM compressors win with pre-trained knowledge. Midicoth closes the gap with pure statistics β€” no mixer, no gradient descent, just counting.
The Tweedie denoising layer adds 2.3–2.7% on every file tested β€” the most consistent component in the ablation. Adding SSE or logistic mixers made things worse. In the online setting, count-based beats gradient-based.
No external dependencies. Fully deterministic. Bit-exact encode/decode. ~60 KB/s throughput.
πŸ’» Code: https://github.com/robtacconelli/midicoth
πŸ“„ Paper: Micro-Diffusion Compression -- Binary Tree Tweedie Denoising for Online Probability Estimation (2603.08771)
⭐ Space: robtacconelli/midicoth

If you ever wondered whether diffusion ideas belong in data compression β€” here's proof they do. ⭐ appreciated!