Picture for Lu Yin

Lu Yin

Mix-LN: Unleashing the Power of Deeper Layers by Combining Pre-LN and Post-LN

Add code
Dec 18, 2024
Viaarxiv icon

Aspect-Based Few-Shot Learning

Add code
Dec 17, 2024
Viaarxiv icon

Condense, Don't Just Prune: Enhancing Efficiency and Performance in MoE Layer Pruning

Add code
Nov 26, 2024
Figure 1 for Condense, Don't Just Prune: Enhancing Efficiency and Performance in MoE Layer Pruning
Figure 2 for Condense, Don't Just Prune: Enhancing Efficiency and Performance in MoE Layer Pruning
Figure 3 for Condense, Don't Just Prune: Enhancing Efficiency and Performance in MoE Layer Pruning
Figure 4 for Condense, Don't Just Prune: Enhancing Efficiency and Performance in MoE Layer Pruning
Viaarxiv icon

Pushing the Limits of Sparsity: A Bag of Tricks for Extreme Pruning

Add code
Nov 21, 2024
Viaarxiv icon

OWLed: Outlier-weighed Layerwise Pruning for Efficient Autonomous Driving Framework

Add code
Nov 12, 2024
Viaarxiv icon

Multimodal Contrastive Learning of Urban Space Representations from POI Data

Add code
Nov 09, 2024
Viaarxiv icon

TODO: Enhancing LLM Alignment with Ternary Preferences

Add code
Nov 02, 2024
Figure 1 for TODO: Enhancing LLM Alignment with Ternary Preferences
Figure 2 for TODO: Enhancing LLM Alignment with Ternary Preferences
Figure 3 for TODO: Enhancing LLM Alignment with Ternary Preferences
Figure 4 for TODO: Enhancing LLM Alignment with Ternary Preferences
Viaarxiv icon

Full-Rank No More: Low-Rank Weight Training for Modern Speech Recognition Models

Add code
Oct 10, 2024
Figure 1 for Full-Rank No More: Low-Rank Weight Training for Modern Speech Recognition Models
Figure 2 for Full-Rank No More: Low-Rank Weight Training for Modern Speech Recognition Models
Figure 3 for Full-Rank No More: Low-Rank Weight Training for Modern Speech Recognition Models
Figure 4 for Full-Rank No More: Low-Rank Weight Training for Modern Speech Recognition Models
Viaarxiv icon

Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning

Add code
Oct 09, 2024
Figure 1 for Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning
Figure 2 for Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning
Figure 3 for Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning
Figure 4 for Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning
Viaarxiv icon

Are Sparse Neural Networks Better Hard Sample Learners?

Add code
Sep 13, 2024
Viaarxiv icon