Picture for Saurabh Agarwal

Saurabh Agarwal

Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition

Add code
Oct 08, 2024
Figure 1 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Figure 2 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Figure 3 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Figure 4 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Viaarxiv icon

LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding

Add code
Apr 29, 2024
Viaarxiv icon

CHAI: Clustered Head Attention for Efficient LLM Inference

Add code
Mar 12, 2024
Viaarxiv icon

Decoding Speculative Decoding

Add code
Feb 02, 2024
Viaarxiv icon

MultiFusionNet: Multilayer Multimodal Fusion of Deep Neural Networks for Chest X-Ray Image Classification

Add code
Jan 01, 2024
Figure 1 for MultiFusionNet: Multilayer Multimodal Fusion of Deep Neural Networks for Chest X-Ray Image Classification
Figure 2 for MultiFusionNet: Multilayer Multimodal Fusion of Deep Neural Networks for Chest X-Ray Image Classification
Figure 3 for MultiFusionNet: Multilayer Multimodal Fusion of Deep Neural Networks for Chest X-Ray Image Classification
Figure 4 for MultiFusionNet: Multilayer Multimodal Fusion of Deep Neural Networks for Chest X-Ray Image Classification
Viaarxiv icon

Cuttlefish: Low-Rank Model Training without All the Tuning

Add code
May 05, 2023
Viaarxiv icon

BagPipe: Accelerating Deep Recommendation Model Training

Add code
Feb 24, 2022
Figure 1 for BagPipe: Accelerating Deep Recommendation Model Training
Figure 2 for BagPipe: Accelerating Deep Recommendation Model Training
Figure 3 for BagPipe: Accelerating Deep Recommendation Model Training
Figure 4 for BagPipe: Accelerating Deep Recommendation Model Training
Viaarxiv icon

Pufferfish: Communication-efficient Models At No Extra Cost

Add code
Mar 05, 2021
Figure 1 for Pufferfish: Communication-efficient Models At No Extra Cost
Figure 2 for Pufferfish: Communication-efficient Models At No Extra Cost
Figure 3 for Pufferfish: Communication-efficient Models At No Extra Cost
Figure 4 for Pufferfish: Communication-efficient Models At No Extra Cost
Viaarxiv icon

On the Utility of Gradient Compression in Distributed Training Systems

Add code
Mar 03, 2021
Figure 1 for On the Utility of Gradient Compression in Distributed Training Systems
Figure 2 for On the Utility of Gradient Compression in Distributed Training Systems
Figure 3 for On the Utility of Gradient Compression in Distributed Training Systems
Figure 4 for On the Utility of Gradient Compression in Distributed Training Systems
Viaarxiv icon

AutoFreeze: Automatically Freezing Model Blocks to Accelerate Fine-tuning

Add code
Feb 02, 2021
Figure 1 for AutoFreeze: Automatically Freezing Model Blocks to Accelerate Fine-tuning
Figure 2 for AutoFreeze: Automatically Freezing Model Blocks to Accelerate Fine-tuning
Figure 3 for AutoFreeze: Automatically Freezing Model Blocks to Accelerate Fine-tuning
Figure 4 for AutoFreeze: Automatically Freezing Model Blocks to Accelerate Fine-tuning
Viaarxiv icon