Picture for Shengen Yan

Shengen Yan

AgentSociety Challenge: Designing LLM Agents for User Modeling and Recommendation on Web Platforms

Add code
Feb 26, 2025
Viaarxiv icon

Megrez-Omni Technical Report

Add code
Feb 19, 2025
Viaarxiv icon

DLFR-VAE: Dynamic Latent Frame Rate VAE for Video Generation

Add code
Feb 17, 2025
Viaarxiv icon

FrameFusion: Combining Similarity and Importance for Video Token Reduction on Large Visual Language Models

Add code
Dec 30, 2024
Figure 1 for FrameFusion: Combining Similarity and Importance for Video Token Reduction on Large Visual Language Models
Viaarxiv icon

MBQ: Modality-Balanced Quantization for Large Vision-Language Models

Add code
Dec 27, 2024
Viaarxiv icon

E-CAR: Efficient Continuous Autoregressive Image Generation via Multistage Modeling

Add code
Dec 19, 2024
Figure 1 for E-CAR: Efficient Continuous Autoregressive Image Generation via Multistage Modeling
Figure 2 for E-CAR: Efficient Continuous Autoregressive Image Generation via Multistage Modeling
Figure 3 for E-CAR: Efficient Continuous Autoregressive Image Generation via Multistage Modeling
Figure 4 for E-CAR: Efficient Continuous Autoregressive Image Generation via Multistage Modeling
Viaarxiv icon

CSKV: Training-Efficient Channel Shrinking for KV Cache in Long-Context Scenarios

Add code
Sep 16, 2024
Figure 1 for CSKV: Training-Efficient Channel Shrinking for KV Cache in Long-Context Scenarios
Figure 2 for CSKV: Training-Efficient Channel Shrinking for KV Cache in Long-Context Scenarios
Figure 3 for CSKV: Training-Efficient Channel Shrinking for KV Cache in Long-Context Scenarios
Figure 4 for CSKV: Training-Efficient Channel Shrinking for KV Cache in Long-Context Scenarios
Viaarxiv icon

Efficient Expert Pruning for Sparse Mixture-of-Experts Language Models: Enhancing Performance and Reducing Inference Costs

Add code
Jul 01, 2024
Viaarxiv icon

MoA: Mixture of Sparse Attention for Automatic Large Language Model Compression

Add code
Jun 21, 2024
Viaarxiv icon

DiTFastAttn: Attention Compression for Diffusion Transformer Models

Add code
Jun 12, 2024
Figure 1 for DiTFastAttn: Attention Compression for Diffusion Transformer Models
Figure 2 for DiTFastAttn: Attention Compression for Diffusion Transformer Models
Figure 3 for DiTFastAttn: Attention Compression for Diffusion Transformer Models
Figure 4 for DiTFastAttn: Attention Compression for Diffusion Transformer Models
Viaarxiv icon