Optimizing Compiler for Neuromorphic Hardware

Master Thesis (2025)
Author(s)

S. Yu (TU Delft - Electrical Engineering, Mathematics and Computer Science)

Contributor(s)

R.K. Bishnoi – Mentor (TU Delft - Computer Engineering)

S.S. Chakraborty – Mentor (TU Delft - Programming Languages)

A. van Deursen – Graduation committee member (TU Delft - Software Engineering)

S.S. Diware – Graduation committee member (TU Delft - Computer Engineering)

Faculty
Electrical Engineering, Mathematics and Computer Science
More Info
expand_more
Publication Year
2025
Language
English
Graduation Date
31-01-2025
Awarding Institution
Delft University of Technology
Programme
['Computer Engineering', 'Electrical Engineering | Embedded Systems']
Faculty
Electrical Engineering, Mathematics and Computer Science
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

The rapid advancement of neural network applications, including multilayer perceptrons (MLP) and deep convolutional neural networks (CNN), has revolutionized domains such as image recognition, speech processing, and classification. However, the increasing depth and complexity of neural network workloads impose significant computational and energy demands. Conventional hardware, such as CPUs and GPUs, faces growing challenges in meeting these demands due to the "memory wall" problem and the diminishing benefits of Moore’s law and Dennard scaling. These issues are particularly pronounced in edge devices where power and energy efficiency are critical.

To address these limitations, computing-in-memory (CIM) architectures, particularly memristor-based crossbar arrays, have emerged as promising solutions. CIM reduces data movement by performing computations directly within memory, significantly improving energy efficiency and performance. Memristor crossbars excel in analog matrix-vector multiplication (MVM), a fundamental operation in neural networks, making them an ideal candidate for accelerating neural network workloads.

This thesis proposes an end-to-end compilation framework that automates the translation and optimization of neural networks for CIM architectures. The framework supports converting high-level MLP models represented in PyTorch into low-level instructions optimized for crossbar-based spatial CIM architectures. Comprehensive experiments explore the impact of various quantization schemes and design space parameters, revealing trade-offs between performance, energy efficiency, and resource utilization. The results demonstrate the framework's potential to support diverse neuromorphic systems and facilitate the efficient deployment of neural networks on CIM architectures.

Files

License info not available
warning

File under embargo until 24-01-2027