What is the difference between network sparsification and model pruning? I watched USENIX ATC '21 - Octo: INT8 Training with Loss-aware Compensation and Backward Quantization for Tiny (at 01:29sec) where they state them as two different methods to simplify the model.
I have difficulties understanding what is the difference.