Title
Studying the Effects of Feature Extraction Settings on the Accuracy and Memory Requirements of Neural Networks for Keyword Spotting
Abstract
Due to the always-on nature of keyword spotting (KWS) systems, low power consumption micro-controller units (MCU) are the best choices as deployment devices. However, small computation power and memory budget of MCUs can harm the accuracy requirements. Although, many studies have been conducted to design small memory footprint neural networks to address this problem, the effects of different feature extraction settings are rarely studied. This work addresses this important question by first, comparing six of the most popular and state of the art neural network architectures for KWS on the Google Speech-Commands dataset. Then, keeping the network architectures unchanged it performs comprehensive investigations on the effects of different frequency transformation settings, such as number of used mel-frequency cepstrum coefficients (MFCCs) and length of the stride window, on the accuracy and memory footprint (RAM/ROM) of the models. The results show different preprocessing settings can change the accuracy and RAM/ROM requirements significantly of the models. Furthermore, it is shown that DS-CNN outperforms the other architectures in terms of accuracy with a value of 93.47% with least amount of ROM requirements, while the GRU outperforms all other networks with an accuracy of 91.02% with smallest RAM requirements.
Year
DOI
Venue
2018
10.1109/ICCE-Berlin.2018.8576243
2018 IEEE 8th International Conference on Consumer Electronics - Berlin (ICCE-Berlin)
Keywords
Field
DocType
memory requirements,keyword spotting systems,KWS,low power consumption microcontroller units,memory footprint neural networks,Google Speech-Commands dataset,RAM/ROM requirements,neural network architectures,frequency transformation settings,feature extraction settings,DS-CNN
Computer science,Cepstrum,Network architecture,Recurrent neural network,Feature extraction,Keyword spotting,Microcontroller,Artificial neural network,Memory footprint,Computer engineering
Conference
ISSN
ISBN
Citations 
2166-6814
978-1-5386-6096-6
1
PageRank 
References 
Authors
0.40
9
5
Name
Order
Citations
PageRank
Muhammad Shahnawaz110.40
Emanuele Plebani232.81
Ivana Guaneri310.40
Danilo Pau4209.74
Marco Marcon55211.10