Picture for Lu Yin

Lu Yin

Condense, Don't Just Prune: Enhancing Efficiency and Performance in MoE Layer Pruning

Add code
Nov 26, 2024
Figure 1 for Condense, Don't Just Prune: Enhancing Efficiency and Performance in MoE Layer Pruning
Figure 2 for Condense, Don't Just Prune: Enhancing Efficiency and Performance in MoE Layer Pruning
Figure 3 for Condense, Don't Just Prune: Enhancing Efficiency and Performance in MoE Layer Pruning
Figure 4 for Condense, Don't Just Prune: Enhancing Efficiency and Performance in MoE Layer Pruning
Viaarxiv icon

Pushing the Limits of Sparsity: A Bag of Tricks for Extreme Pruning

Add code
Nov 21, 2024
Viaarxiv icon

OWLed: Outlier-weighed Layerwise Pruning for Efficient Autonomous Driving Framework

Add code
Nov 12, 2024
Viaarxiv icon

Multimodal Contrastive Learning of Urban Space Representations from POI Data

Add code
Nov 09, 2024
Viaarxiv icon

TODO: Enhancing LLM Alignment with Ternary Preferences

Add code
Nov 02, 2024
Figure 1 for TODO: Enhancing LLM Alignment with Ternary Preferences
Figure 2 for TODO: Enhancing LLM Alignment with Ternary Preferences
Figure 3 for TODO: Enhancing LLM Alignment with Ternary Preferences
Figure 4 for TODO: Enhancing LLM Alignment with Ternary Preferences
Viaarxiv icon

Full-Rank No More: Low-Rank Weight Training for Modern Speech Recognition Models

Add code
Oct 10, 2024
Figure 1 for Full-Rank No More: Low-Rank Weight Training for Modern Speech Recognition Models
Figure 2 for Full-Rank No More: Low-Rank Weight Training for Modern Speech Recognition Models
Figure 3 for Full-Rank No More: Low-Rank Weight Training for Modern Speech Recognition Models
Figure 4 for Full-Rank No More: Low-Rank Weight Training for Modern Speech Recognition Models
Viaarxiv icon

Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning

Add code
Oct 09, 2024
Figure 1 for Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning
Figure 2 for Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning
Figure 3 for Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning
Figure 4 for Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning
Viaarxiv icon

Are Sparse Neural Networks Better Hard Sample Learners?

Add code
Sep 13, 2024
Viaarxiv icon

Robust Active Learning (RoAL): Countering Dynamic Adversaries in Active Learning with Elastic Weight Consolidation

Add code
Aug 15, 2024
Figure 1 for Robust Active Learning (RoAL): Countering Dynamic Adversaries in Active Learning with Elastic Weight Consolidation
Figure 2 for Robust Active Learning (RoAL): Countering Dynamic Adversaries in Active Learning with Elastic Weight Consolidation
Figure 3 for Robust Active Learning (RoAL): Countering Dynamic Adversaries in Active Learning with Elastic Weight Consolidation
Figure 4 for Robust Active Learning (RoAL): Countering Dynamic Adversaries in Active Learning with Elastic Weight Consolidation
Viaarxiv icon

From GaLore to WeLore: How Low-Rank Weights Non-uniformly Emerge from Low-Rank Gradients

Add code
Jul 15, 2024
Figure 1 for From GaLore to WeLore: How Low-Rank Weights Non-uniformly Emerge from Low-Rank Gradients
Figure 2 for From GaLore to WeLore: How Low-Rank Weights Non-uniformly Emerge from Low-Rank Gradients
Figure 3 for From GaLore to WeLore: How Low-Rank Weights Non-uniformly Emerge from Low-Rank Gradients
Figure 4 for From GaLore to WeLore: How Low-Rank Weights Non-uniformly Emerge from Low-Rank Gradients
Viaarxiv icon