Rethinking the value of network pruning

Tags: paper ml
State: None
Code: None


  • States LTH's method of finding a winning ticket is not necessary
  • It's commonly believed in literature that you have to: train, then prune and then fine-tune
  • But this is not necessary for unstructured pruning
  • Instead, you can find a better set of hparams for the optimization method, e.g. for SGD
  • For unstructured pruning this does not scale to ImageNet. They do not know why. Only comparable perf for smaller datasets (and models?)