Picture for Peng Jin

Peng Jin

Senior member, IEEE

MoH: Multi-Head Attention as Mixture-of-Head Attention

Add code
Oct 15, 2024
Viaarxiv icon

MoE++: Accelerating Mixture-of-Experts Methods with Zero-Computation Experts

Add code
Oct 09, 2024
Viaarxiv icon

MUSE: Mamba is Efficient Multi-scale Learner for Text-video Retrieval

Add code
Aug 20, 2024
Figure 1 for MUSE: Mamba is Efficient Multi-scale Learner for Text-video Retrieval
Figure 2 for MUSE: Mamba is Efficient Multi-scale Learner for Text-video Retrieval
Figure 3 for MUSE: Mamba is Efficient Multi-scale Learner for Text-video Retrieval
Figure 4 for MUSE: Mamba is Efficient Multi-scale Learner for Text-video Retrieval
Viaarxiv icon

Local Action-Guided Motion Diffusion Model for Text-to-Motion Generation

Add code
Jul 15, 2024
Viaarxiv icon

LOOK-M: Look-Once Optimization in KV Cache for Efficient Multimodal Long-Context Inference

Add code
Jun 26, 2024
Viaarxiv icon

RAP: Efficient Text-Video Retrieval with Sparse-and-Correlated Adapter

Add code
May 29, 2024
Viaarxiv icon

LLMBind: A Unified Modality-Task Integration Framework

Add code
Mar 08, 2024
Viaarxiv icon

SPHINX-X: Scaling Data and Parameters for a Family of Multi-modal Large Language Models

Add code
Feb 08, 2024
Viaarxiv icon

MoE-LLaVA: Mixture of Experts for Large Vision-Language Models

Add code
Feb 04, 2024
Figure 1 for MoE-LLaVA: Mixture of Experts for Large Vision-Language Models
Figure 2 for MoE-LLaVA: Mixture of Experts for Large Vision-Language Models
Figure 3 for MoE-LLaVA: Mixture of Experts for Large Vision-Language Models
Figure 4 for MoE-LLaVA: Mixture of Experts for Large Vision-Language Models
Viaarxiv icon

Instance Brownian Bridge as Texts for Open-vocabulary Video Instance Segmentation

Add code
Jan 18, 2024
Viaarxiv icon