Picture for Huiyin Xue

Huiyin Xue

Pit One Against Many: Leveraging Attention-head Embeddings for Parameter-efficient Multi-head Attention

Add code
Oct 11, 2023
Viaarxiv icon

HashFormers: Towards Vocabulary-independent Pre-trained Transformers

Add code
Oct 14, 2022
Figure 1 for HashFormers: Towards Vocabulary-independent Pre-trained Transformers
Figure 2 for HashFormers: Towards Vocabulary-independent Pre-trained Transformers
Figure 3 for HashFormers: Towards Vocabulary-independent Pre-trained Transformers
Figure 4 for HashFormers: Towards Vocabulary-independent Pre-trained Transformers
Viaarxiv icon