Offline Compression of Convolutional Neural Networks on Edge Devices

Bachelor Thesis (2020)
Author(s)

Simon Tulling (TU Delft - Electrical Engineering, Mathematics and Computer Science)

Contributor(s)

Lydia Chen – Mentor (TU Delft - Data-Intensive Systems)

Amirmasoud Ghiassi – Graduation committee member (TU Delft - Data-Intensive Systems)

Bart Cox – Graduation committee member (TU Delft - Electrical Engineering, Mathematics and Computer Science)

Marco Zuñiga Zamalloa – Coach (TU Delft - Embedded Systems)

More Info
expand_more
Publication Year
2020
Language
English
Graduation Date
22-06-2020
Awarding Institution
Project
CSE3000 Research Project
Programme
Computer Science and Engineering
Downloads counter
181
Collections
thesis
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Edge Devices and Artificial Intelligence are important and ever increasing fields in technology. Yet their combination is lacking because the neural networks used in AI are being made increasingly large and complex while edge devices lack the resources to keep up with these developments. Neural network model compression will allow these edge devices to run these models due to overcoming memory constraints. This paper proposes to use both singular value decomposition and canonical polyadic decomposition as a way to decrease the size of convolutional neural networks at the cost of some accuracy. This compression pipeline can be run on an edge device and is configurable to change the trade-off between file size and accuracy. This creates a possibility to run convolutional neural networks natively on edge devices.

Files

Research_8_.pdf
(pdf | 0.622 Mb)
License info not available