Demystifying Node Pruning through the Perspective of Subspaces

Keywords

Loading...
Thumbnail Image

Issue Date

2024-08-01

Language

en

Document type

Journal Title

Journal ISSN

Volume Title

Publisher

Title

ISSN

Volume

Issue

Startpage

Endpage

DOI

Abstract

A significant increase in the commercial use of deep neural network models increases the need for efficient AI. Node pruning is the art of removing computational nodes such as neurons, filters, attention heads, or even entire layers, such that the network performance is kept at a maximum after retraining. This can significantly reduce the inference time of a deep network and thus enhance its efficiency. Previous works have exploited the ability to recover performance by reorganizing network parameters while pruning via linear least squares approximation of the original neural responses. In this work, we relate this method to projecting node activity to an orthogonal subspace, such that pruning within the subspace translates to effective node pruning. We find that this reconstruction barely affects network performance compared to simply pruning units after retraining. However, we use these findings to develop novel layer-wise importance scoring methods and the relation of such layer-wise rankings across layers within a network to formulate a measure of global unit importance. Through application to VGG-16 trained on the ILSVRC dataset, we show that this global unit importance outperforms the baseline methods tested herein.

Description

Citation

Faculty

Faculteit der Sociale Wetenschappen