Print Email Facebook Twitter Analysis of the effect of caching convolutional network layers on resource constraint devices Title Analysis of the effect of caching convolutional network layers on resource constraint devices Author van Lil, Wouter (TU Delft Electrical Engineering, Mathematics and Computer Science) Contributor Chen, Lydia Y. (mentor) Cox, B.A. (graduation committee) Ghiassi, S. (graduation committee) Degree granting institution Delft University of Technology Programme Computer Science and Engineering Project Optimizing Multiple Deep Learning Models on Edge Devices Date 2020-06-22 Abstract Using transfer learning, convolutional neural networks for different purposes can have similar layers which can be reused by caching them, reducing their load time. Four ways of loading and executing these layers, bulk, linear, DeepEye and partial loading, were analysed under different memory constraints and different amounts of similar networks. When there is sufficient memory, caching will decrease the loading time and will always influence the single threaded bulk and linear mode. On the multithreaded approaches this only holds when the loading time is longer than the execution time. This depends largely on what network will be run. When memory constraints are applied caching can be a way to still run multiple networks without much increased cost. It can also be opted to use less memory on a device and use transfer learning with caching to still get the same results. Subject deep learningtransfer learningcachingpartial loading To reference this document use: http://resolver.tudelft.nl/uuid:85251e58-77d6-40e5-b77a-06cc1a7798d1 Part of collection Student theses Document type bachelor thesis Rights © 2020 Wouter van Lil Files PDF Scriptie_wouter.pdf 236.38 KB Close viewer /islandora/object/uuid:85251e58-77d6-40e5-b77a-06cc1a7798d1/datastream/OBJ/view