Picture for Tong Zhu

Tong Zhu

Chain-of-Tools: Utilizing Massive Unseen Tools in the CoT Reasoning of Frozen Language Models

Add code
Mar 21, 2025
Viaarxiv icon

Linear-MoE: Linear Sequence Modeling Meets Mixture-of-Experts

Add code
Mar 07, 2025
Viaarxiv icon

Iterative Value Function Optimization for Guided Decoding

Add code
Mar 05, 2025
Viaarxiv icon

LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training

Add code
Nov 24, 2024
Viaarxiv icon

NesTools: A Dataset for Evaluating Nested Tool Learning Abilities of Large Language Models

Add code
Oct 15, 2024
Viaarxiv icon

CLIP-MoE: Towards Building Mixture of Experts for CLIP with Diversified Multiplet Upcycling

Add code
Sep 28, 2024
Viaarxiv icon

ConflictBank: A Benchmark for Evaluating the Influence of Knowledge Conflicts in LLM

Add code
Aug 22, 2024
Viaarxiv icon

Learning to Refuse: Towards Mitigating Privacy Risks in LLMs

Add code
Jul 14, 2024
Viaarxiv icon

LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training

Add code
Jun 24, 2024
Viaarxiv icon

Timo: Towards Better Temporal Reasoning for Language Models

Add code
Jun 20, 2024
Viaarxiv icon