@BerivanISIK
Berivan Isik
2 years
@miniapeur There is an (not very tight) upper bound on the output distortion when pruning a single connection that helps with adjusting layer-wise sparsity in a greedy manner:
3
0
8

Replies

@miniapeur
Mathieu Alain
2 years
How should sparsity be promoted in a neural network? First layers, last layers, uniformly? Any theoretical results about this?
4
1
24
@miniapeur
Mathieu Alain
2 years
0
0
1
@BerivanISIK
Berivan Isik
2 years
@miniapeur We extend this to a more general case beyond single-connection pruning in Theorem 1 here:
2
0
3
@BerivanISIK
Berivan Isik
2 years
@miniapeur There is also by @utkuevci that empirically compares different sparsity distributions.
1
0
3