TY - JOUR T1 - Automatic Convolution Neural Network Model Compression Framework for Resource-Constrained Embedded Systems AU - Jeong, Jonghun AU - Lee, Dasom AU - Jung, Hyeonseok AU - Yang, Hoeseok JO - Journal of KIISE, JOK PY - 2020 DA - 2020/1/14 DO - 10.5626/JOK.2020.47.2.136 KW - internet-of-things device KW - embedded system KW - microprocessor KW - neural network KW - neural network compression KW - filter pruning AB - Recently, attempts have been made to directly execute various convolutional neural network applications in resource-constrained embedded systems such as IoT. However, since embedded systems have limited computational capability and memory, the size of the neural network model that can be executed is restricted and may not satisfy real-time constraints. Therefore, in this paper, we propose a framework that automatically compresses a given neural network model to satisfy memory and execution time requirements and automatically generates code that can be executed on the target embedded system. Using the proposed framework, we demonstrate that the given neural network models can be automatically optimized for two STM32 Nucleo series boards with different HW specifications for various execution time and memory requirements.