DyPrune: Dynamic Pruning Rates for Neural Networks Artigo de Conferência Capítulo de livro uri icon

resumo

  • Neural networks have achieved remarkable success in various applications such as image classification, speech recognition, and natural language processing. However, the growing size of neural networks poses significant challenges in terms of memory usage, computational cost, and deployment on resource-constrained devices. Pruning is a popular technique to reduce the complexity of neural networks by removing unnecessary connections, neurons, or filters. In this paper, we present novel pruning algorithms that can reduce the number of parameters in neural networks by up to 98% without sacrificing accuracy. This is done by scaling the pruning rate of the models to the size of the model and scheduling the pruning to execute throughout the training of the model. Code related to this work is openly available.

autores

  • Jonker, Richard Adolph Aires
  • Poudel, Roshan
  • Fajarda, Olga
  • José Luís Oliveira
  • Matos, Sergio

data de publicação

  • 2023