Tag: On-device

AI Machine Learning & Data Science Research

Google’s 1.3 MiB On-Device Model Brings High-Performance Disfluency Detection Down to Size

A research team from Google Research proposes small, fast, on-device disfluency detection models based on the BERT architecture. The smallest model size is only 1.3 MiB, representing a size reduction of two orders of magnitude and an inference latency reduction of a factor of eight compared to state-of-the-art BERT-based models.