Hasty Briefsbeta

Bilingual

SMLL: Using 200MB of Neural Network to Save 400 Bytes

3 months ago
  • #neural-networks
  • #machine-learning
  • #compression
  • SMLL compression uses a 200MB neural network to achieve 10x better compression than gzip on text.
  • The method combines LLMs with arithmetic coding to approach the theoretical limit of compression.
  • Compression ratios vary by content type, with LLM-generated text achieving the best results (14.96x).
  • Compression improves with text length due to better context accumulation by the LLM.
  • SMLL is significantly slower than gzip (10,000x slower) due to the computational cost of neural network inference.
  • The trade-off between model size, speed, and compression efficiency depends on specific use cases.
  • The connection between compression and intelligence is highlighted, with compression efficiency reflecting model perplexity.
  • Practical applications include scenarios where storage costs outweigh compute costs, but not for high-speed needs like HTTP responses.
  • Future work could explore whether LLMs outperform simple lookup tables on novel text, further investigating the compression-intelligence link.