Hasty Briefsbeta

Bilingual

What even is a small language model now?

a year ago
  • #AI
  • #Machine Learning
  • #Small Models
  • The definition of 'small models' has evolved from a few million parameters in 2018 to 30B-70B parameters today, now considered 'small' if they run on a single GPU.
  • Small models are categorized into two types: those optimized for mobile/edge devices and those requiring a single GPU.
  • Small models are often narrow and optimized for specific tasks, offering advantages like lower costs, faster inference, and better privacy.
  • The bar for what's considered 'small' keeps shifting, with quantization and engineering enabling even 70B models to run on consumer GPUs.
  • Some long-standing small models, like Google Translate and AWS Textract, continue to be widely used despite not being cutting-edge.
  • Small models are gaining importance due to their efficiency, cost-effectiveness, and ability to compete with larger models like GPT-3.5 in benchmarks.