Treffer: One-cycle pruning: pruning convnets with tight training budget

Title:
One-cycle pruning: pruning convnets with tight training budget
Contributors:
Institut Polytechnique de Paris (IP Paris), Département Advanced Research And Techniques For Multidimensional Imaging Systems (TSP - ARTEMIS), Institut Mines-Télécom [Paris] (IMT)-Télécom SudParis (TSP), Institut Mines-Télécom [Paris] (IMT)-Institut Polytechnique de Paris (IP Paris)-Institut Polytechnique de Paris (IP Paris), ARMEDIA (ARMEDIA-SAMOVAR), Services répartis, Architectures, MOdélisation, Validation, Administration des Réseaux (SAMOVAR), Institut Mines-Télécom [Paris] (IMT)-Institut Polytechnique de Paris (IP Paris)-Institut Polytechnique de Paris (IP Paris)-Institut Mines-Télécom [Paris] (IMT)-Télécom SudParis (TSP), Université de Mons = University of Mons (UMONS)
Source:
2022 IEEE International Conference on Image Processing (ICIP). :4128-4132
Publisher Information:
CCSD; IEEE, 2022.
Publication Year:
2022
Collection:
collection:TELECOM-SUDPARIS
collection:IP_PARIS
collection:INSTITUTS-TELECOM
collection:INSTITUT-MINES-TELECOM
collection:DEPARTEMENT-DE-MATHEMATIQUES
collection:IP-PARIS-MATHEMATIQUES
collection:IP-PARIS-INFORMATION-COMMUNICATION-ELECTRONIQUE
collection:IP-PARIS-INFORMATIQUE-DONNEES-ET-IA
collection:SAMOVAR
Subject Geographic:
Original Identifier:
HAL: hal-03930230
Document Type:
Konferenz conferenceObject<br />Conference papers
Language:
English
ISBN:
978-1-66549-621-6
Relation:
info:eu-repo/semantics/altIdentifier/doi/10.1109/ICIP46576.2022.9897980
DOI:
10.1109/ICIP46576.2022.9897980
Accession Number:
edshal.hal.03930230v1
Database:
HAL

Weitere Informationen

Introducing sparsity in a convnet has been an efficient way to reduce its complexity while keeping its performance almost intact. Most of the time, sparsity is introduced using a three-stage pipeline: 1) training the model to convergence, 2) pruning the model, 3) fine-tuning the pruned model to recover performance. The last two steps are often performed iteratively, leading to reasonable results but also to a time-consuming process. In our work, we propose to remove the first step of the pipeline and to combine the two others in a single training-pruning cycle, allowing the model to jointly learn the optimal weights while being pruned. We do this by introducing a novel pruning schedule, named One-Cycle Pruning (OCP), which starts pruning from the beginning of the training, and until its very end. Experiments conducted on a variety of combinations between architectures (VGG-16, ResNet-18), datasets (CIFAR-10, CIFAR-100, Caltech-101), and sparsity values (80%, 90%, 95%) show that not only OCP consistently outperforms common pruning schedules such as One-Shot, Iterative and Automated Gradual Pruning, but also that it drastically reduces the required training budget. More-over, experiments following the Lottery Ticket Hypothesis show that OCP allows to find higher quality and more stable pruned networks.