MN

M. Naderan-Tahan

info

Please Note

2 records found

Towards Programmable Computing-in-Memory Accelerators

Design of a general purpose programming model

For a very long period of time, computing could meet the increasing demands of different applications due to the continued downscaling of transistors, which allowed data to be processed at a higher frequency. In the early 2000s, predictions about the physical limits and rising co ...
Model compression techniques are crucial for reducing the deployment cost of large neural networks. Among these, depth pruning (removing layers/blocks) and width pruning (removing sections within layers) are essential for reducing memory footprint and inference latency. While var ...