Picture for Piotr Nawrot

Piotr Nawrot

Dynamic Memory Compression: Retrofitting LLMs for Accelerated Inference

Add code
Mar 14, 2024
Viaarxiv icon

nanoT5: A PyTorch Framework for Pre-training and Fine-tuning T5-style Models with Limited Resources

Add code
Sep 05, 2023
Viaarxiv icon

No Train No Gain: Revisiting Efficient Training Algorithms For Transformer-based Language Models

Add code
Jul 26, 2023
Viaarxiv icon

Efficient Transformers with Dynamic Token Pooling

Add code
Nov 17, 2022
Viaarxiv icon

Hierarchical Transformers Are More Efficient Language Models

Add code
Oct 26, 2021
Figure 1 for Hierarchical Transformers Are More Efficient Language Models
Figure 2 for Hierarchical Transformers Are More Efficient Language Models
Figure 3 for Hierarchical Transformers Are More Efficient Language Models
Figure 4 for Hierarchical Transformers Are More Efficient Language Models
Viaarxiv icon