<scp>Tailor</scp> : Altering Skip Connections for Resource-Efficient Inference
نویسندگان
چکیده
Deep neural networks use skip connections to improve training convergence. However, these are costly in hardware, requiring extra buffers and increasing on- off-chip memory utilization bandwidth requirements. In this paper, we show that can be optimized for hardware when tackled with a hardware-software codesign approach. We argue while network’s needed the network learn, they later removed or shortened provide more efficient implementation minimal no accuracy loss. introduce Tailor , tool whose hardware-aware algorithm gradually removes shortens fully trained lower their cost. improves resource by up 34% BRAMs, 13% FFs, 16% LUTs on-chip, dataflow-style architectures. increases performance 30% reduces 45% 2D processing element array architecture.
منابع مشابه
Skip Connections Eliminate Singularities
Skip connections made the training of very deep networks possible and have become an indispensable component in a variety of neural architectures. A completely satisfactory explanation for their success remains elusive. Here, we present a novel explanation for the benefits of skip connections in training very deep networks. The difficulty of training deep networks is partly due to the singulari...
متن کاملSkip Connections Eliminate Singularities
Skip connections made the training of very deep networks possible and have become an indispensable component in a variety of neural architectures. A completely satisfactory explanation for their success remains elusive. Here, we present a novel explanation for the benefits of skip connections in training very deep networks. The difficulty of training deep networks is partly due to the singulari...
متن کاملMonaural Singing Voice Separation with Skip-Filtering Connections and Recurrent Inference of Time-Frequency Mask
Singing voice separation based on deep learning relies on the usage of time-frequency masking. In many cases the masking process is not a learnable function or is not encapsulated into the deep learning optimization. Consequently, most of the existing methods rely on a post processing step using the generalized Wiener filtering. This work proposes a method that learns and optimizes (during trai...
متن کاملMultiple Skip Connections and Dilated Convolutions for Semantic Segmentation
We propose a scale-aware semantic segmentation method specifically for small objects. The contributions of this method are 1) to feed the features of a small region by using multiple skip connections, and 2) to extract context from multiple receptive fields by using multiple dilated convolution blocks. The proposed method has achieved high accuracy in the Cityscapes dataset. In comparison with ...
متن کاملBeyond Skip Connections: Top-Down Modulation for Object Detection
In recent years, we have seen tremendous progress in the field of object detection. Most of the recent improvements have been achieved by targeting deeper feedforward networks. However, many hard object categories such as bottle, remote, etc. require representation of fine details and not just coarse, semantic representations. But most of these fine details are lost in the early convolutional l...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: ACM Transactions on Reconfigurable Technology and Systems
سال: 2023
ISSN: ['1936-7414', '1936-7406']
DOI: https://doi.org/10.1145/3624990