Hasty Briefsbeta

New Apple Study Shows LLMs Can Tell What You're Doing from Audio and Motion Data

8 hours ago
  • #Activity Recognition
  • #LLM
  • #Sensor Data
  • Apple researchers explore using LLMs to analyze audio and motion data for better user activity recognition.
  • LLMs show potential in improving activity analysis precision, even with limited sensor data.
  • Study uses Ego4D dataset, focusing on 12 diverse activities like household tasks and sports.
  • LLMs achieve significant accuracy in zero- and one-shot classification without task-specific training.
  • Combining multiple models enhances activity and health data analysis, especially with insufficient raw sensor data.
  • Apple provides supplemental materials for researchers to reproduce the study's results.