Picture for Mingjie Sun

Mingjie Sun

Stronger Normalization-Free Transformers

Add code
Dec 11, 2025
Viaarxiv icon

Align$^3$GR: Unified Multi-Level Alignment for LLM-based Generative Recommendation

Add code
Nov 14, 2025
Viaarxiv icon

Command-V: Pasting LLM Behaviors via Activation Profiles

Add code
Jun 23, 2025
Viaarxiv icon

Noise-Consistent Siamese-Diffusion for Medical Image Synthesis and Segmentation

Add code
May 09, 2025
Viaarxiv icon

ConSense: Continually Sensing Human Activity with WiFi via Growing and Picking

Add code
Feb 18, 2025
Viaarxiv icon

Idiosyncrasies in Large Language Models

Add code
Feb 17, 2025
Figure 1 for Idiosyncrasies in Large Language Models
Figure 2 for Idiosyncrasies in Large Language Models
Figure 3 for Idiosyncrasies in Large Language Models
Figure 4 for Idiosyncrasies in Large Language Models
Viaarxiv icon

Massive Values in Self-Attention Modules are the Key to Contextual Knowledge Understanding

Add code
Feb 03, 2025
Viaarxiv icon

SweetTokenizer: Semantic-Aware Spatial-Temporal Tokenizer for Compact Visual Discretization

Add code
Dec 17, 2024
Figure 1 for SweetTokenizer: Semantic-Aware Spatial-Temporal Tokenizer for Compact Visual Discretization
Figure 2 for SweetTokenizer: Semantic-Aware Spatial-Temporal Tokenizer for Compact Visual Discretization
Figure 3 for SweetTokenizer: Semantic-Aware Spatial-Temporal Tokenizer for Compact Visual Discretization
Figure 4 for SweetTokenizer: Semantic-Aware Spatial-Temporal Tokenizer for Compact Visual Discretization
Viaarxiv icon

Bi-Mamba: Towards Accurate 1-Bit State Space Models

Add code
Nov 18, 2024
Viaarxiv icon

FBI-LLM: Scaling Up Fully Binarized LLMs from Scratch via Autoregressive Distillation

Add code
Jul 09, 2024
Figure 1 for FBI-LLM: Scaling Up Fully Binarized LLMs from Scratch via Autoregressive Distillation
Figure 2 for FBI-LLM: Scaling Up Fully Binarized LLMs from Scratch via Autoregressive Distillation
Figure 3 for FBI-LLM: Scaling Up Fully Binarized LLMs from Scratch via Autoregressive Distillation
Figure 4 for FBI-LLM: Scaling Up Fully Binarized LLMs from Scratch via Autoregressive Distillation
Viaarxiv icon