Picture for Zhouhan Lin

Zhouhan Lin

VQKV: High-Fidelity and High-Ratio Cache Compression via Vector-Quantization

Add code
Mar 17, 2026
Viaarxiv icon

Early Quantization Shrinks Codebook: A Simple Fix for Diversity-Preserving Tokenization

Add code
Mar 17, 2026
Viaarxiv icon

CoDAR: Continuous Diffusion Language Models are More Powerful Than You Think

Add code
Mar 03, 2026
Viaarxiv icon

AdaPonderLM: Gated Pondering Language Models with Token-Wise Adaptive Depth

Add code
Mar 02, 2026
Viaarxiv icon

PonderLM-3: Adaptive Token-Wise Pondering with Differentiable Masking

Add code
Mar 02, 2026
Viaarxiv icon

Towards Compressive and Scalable Recurrent Memory

Add code
Feb 11, 2026
Viaarxiv icon

Flow of Spans: Generalizing Language Models to Dynamic Span-Vocabulary via GFlowNets

Add code
Feb 11, 2026
Viaarxiv icon

Pretraining with Token-Level Adaptive Latent Chain-of-Thought

Add code
Feb 09, 2026
Viaarxiv icon

Next Concept Prediction in Discrete Latent Space Leads to Stronger Language Models

Add code
Feb 09, 2026
Viaarxiv icon

Controlling Exploration-Exploitation in GFlowNets via Markov Chain Perspectives

Add code
Feb 03, 2026
Viaarxiv icon