Picture for Yingyan Celine Lin

Yingyan Celine Lin

MixGCN: Scalable GCN Training by Mixture of Parallelism and Mixture of Accelerators

Add code
Jan 06, 2025
Figure 1 for MixGCN: Scalable GCN Training by Mixture of Parallelism and Mixture of Accelerators
Figure 2 for MixGCN: Scalable GCN Training by Mixture of Parallelism and Mixture of Accelerators
Figure 3 for MixGCN: Scalable GCN Training by Mixture of Parallelism and Mixture of Accelerators
Figure 4 for MixGCN: Scalable GCN Training by Mixture of Parallelism and Mixture of Accelerators
Viaarxiv icon

Layer- and Timestep-Adaptive Differentiable Token Compression Ratios for Efficient Diffusion Transformers

Add code
Dec 22, 2024
Viaarxiv icon

AmoebaLLM: Constructing Any-Shape Large Language Models for Efficient and Instant Deployment

Add code
Nov 15, 2024
Viaarxiv icon

EDGE-LLM: Enabling Efficient Large Language Model Adaptation on Edge Devices via Layerwise Unified Compression and Adaptive Layer Tuning and Voting

Add code
Jun 22, 2024
Viaarxiv icon

Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration

Add code
Jun 22, 2024
Figure 1 for Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration
Figure 2 for Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration
Figure 3 for Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration
Figure 4 for Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration
Viaarxiv icon