Layer-wise Pruning of Transformer Attention Heads for Efficient Language Modeling

Add code
Oct 07, 2021
Figure 1 for Layer-wise Pruning of Transformer Attention Heads for Efficient Language Modeling
Figure 2 for Layer-wise Pruning of Transformer Attention Heads for Efficient Language Modeling
Figure 3 for Layer-wise Pruning of Transformer Attention Heads for Efficient Language Modeling
Figure 4 for Layer-wise Pruning of Transformer Attention Heads for Efficient Language Modeling

Share this with someone who'll enjoy it:

View paper onarxiv icon

Share this with someone who'll enjoy it: