Picture for Ziang Song

Ziang Song

MoE-Pruner: Pruning Mixture-of-Experts Large Language Model using the Hints from Its Router

Add code
Oct 15, 2024
Figure 1 for MoE-Pruner: Pruning Mixture-of-Experts Large Language Model using the Hints from Its Router
Figure 2 for MoE-Pruner: Pruning Mixture-of-Experts Large Language Model using the Hints from Its Router
Figure 3 for MoE-Pruner: Pruning Mixture-of-Experts Large Language Model using the Hints from Its Router
Figure 4 for MoE-Pruner: Pruning Mixture-of-Experts Large Language Model using the Hints from Its Router
Viaarxiv icon

Reward Collapse in Aligning Large Language Models

Add code
May 28, 2023
Viaarxiv icon

Efficient $Φ$-Regret Minimization in Extensive-Form Games via Online Mirror Descent

Add code
Jun 02, 2022
Viaarxiv icon

Sample-Efficient Learning of Correlated Equilibria in Extensive-Form Games

Add code
May 15, 2022
Viaarxiv icon

Learn To Remember: Transformer with Recurrent Memory for Document-Level Machine Translation

Add code
May 03, 2022
Figure 1 for Learn To Remember: Transformer with Recurrent Memory for Document-Level Machine Translation
Figure 2 for Learn To Remember: Transformer with Recurrent Memory for Document-Level Machine Translation
Figure 3 for Learn To Remember: Transformer with Recurrent Memory for Document-Level Machine Translation
Figure 4 for Learn To Remember: Transformer with Recurrent Memory for Document-Level Machine Translation
Viaarxiv icon

When Can We Learn General-Sum Markov Games with a Large Number of Players Sample-Efficiently?

Add code
Oct 08, 2021
Figure 1 for When Can We Learn General-Sum Markov Games with a Large Number of Players Sample-Efficiently?
Viaarxiv icon