Linearly Replaceable Filters for Deep Network Channel Pruning

نویسندگان

چکیده

Convolutional neural networks (CNNs) have achieved remarkable results; however, despite the development of deep learning, practical user applications are fairly limited because heavy can be used solely with latest hardware and software supports. Therefore, network pruning is gaining attention for general in various fields. This paper proposes a novel channel method, Linearly Replaceable Filter (LRF), which suggests that filter approximated by linear combination other filters replaceable. Moreover, an additional method called Weights Compensation proposed to support LRF method. technique effectively reduces output difference caused removing via direct weight modification. Through experiments, we confirmed our achieves state-of-the-art performance several benchmarks. In particular, on ImageNet, LRF-60 approximately 56% FLOPs ResNet-50 without top-5 accuracy drop. Further, through extensive analyses, proved effectiveness approaches.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Automated Pruning for Deep Neural Network Compression

In this work we present a method to improve the pruning step of the current state-of-the-art methodology to compress neural networks. The novelty of the proposed pruning technique is in its differentiability, which allows pruning to be performed during the backpropagation phase of the network training. This enables an end-to-end learning and strongly reduces the training time. The technique is ...

متن کامل

Pruning Filters for Efficient ConvNets

Convolutional Neural Networks (CNNs) are extensively used in image and video recognition, natural language processing and other machine learning applications. The success of CNNs in these areas corresponds with a significant increase in the number of parameters and computation costs. Recent approaches towards reducing these overheads involve pruning and compressing the weights of various layers...

متن کامل

Automatic Tuning for Linearly Tunable Filters

Automatic Tuning for Linearly Tunable Filters. (May 2004) Sung-Ling Huang, B.S., Tatung Institute of Technology, Taiwan; M.S., Tatung Institute of Technology, Taiwan Chair of Advisory Committee: Dr. Aydin I. Karsilayan A new tuning scheme for linearly tunable high-Q filters is proposed. The tuning method is based on using the phase information for both frequency and Q factor tuning. There is no...

متن کامل

Structured Deep Neural Network Pruning via Matrix Pivoting

Deep Neural Networks (DNNs) are the key to the state-of-the-art machine vision, sensor fusion and audio/video signal processing. Unfortunately, their computation complexity and tight resource constraints on the Edge make them hard to leverage on mobile, embedded and IoT devices. Due to great diversity of Edge devices, DNN designers have to take into account the hardware platform and application...

متن کامل

A Deep Neural Network Compression Pipeline: Pruning, Quantization, Huffman Encoding

Neural networks are both computationally intensive and memory intensive, making them difficult to deploy on embedded systems with limited hardware resources. To address this limitation, We introduce a three stage pipeline: pruning, quantization and Huffman encoding, that work together to reduce the storage requirement of neural networks by 35× to 49× without affecting their accuracy. Our method...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2021

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v35i9.16978