Picture for Kang Zhao

Kang Zhao

EgoVid-5M: A Large-Scale Video-Action Dataset for Egocentric Video Generation

Add code
Nov 13, 2024
Figure 1 for EgoVid-5M: A Large-Scale Video-Action Dataset for Egocentric Video Generation
Figure 2 for EgoVid-5M: A Large-Scale Video-Action Dataset for Egocentric Video Generation
Figure 3 for EgoVid-5M: A Large-Scale Video-Action Dataset for Egocentric Video Generation
Figure 4 for EgoVid-5M: A Large-Scale Video-Action Dataset for Egocentric Video Generation
Viaarxiv icon

FastAttention: Extend FlashAttention2 to NPUs and Low-resource GPUs

Add code
Oct 22, 2024
Viaarxiv icon

Beyond 2:4: exploring V:N:M sparsity for efficient transformer inference on GPUs

Add code
Oct 21, 2024
Figure 1 for Beyond 2:4: exploring V:N:M sparsity for efficient transformer inference on GPUs
Figure 2 for Beyond 2:4: exploring V:N:M sparsity for efficient transformer inference on GPUs
Figure 3 for Beyond 2:4: exploring V:N:M sparsity for efficient transformer inference on GPUs
Figure 4 for Beyond 2:4: exploring V:N:M sparsity for efficient transformer inference on GPUs
Viaarxiv icon

FlatQuant: Flatness Matters for LLM Quantization

Add code
Oct 12, 2024
Figure 1 for FlatQuant: Flatness Matters for LLM Quantization
Figure 2 for FlatQuant: Flatness Matters for LLM Quantization
Figure 3 for FlatQuant: Flatness Matters for LLM Quantization
Figure 4 for FlatQuant: Flatness Matters for LLM Quantization
Viaarxiv icon

FreeMask: Rethinking the Importance of Attention Masks for Zero-Shot Video Editing

Add code
Sep 30, 2024
Figure 1 for FreeMask: Rethinking the Importance of Attention Masks for Zero-Shot Video Editing
Figure 2 for FreeMask: Rethinking the Importance of Attention Masks for Zero-Shot Video Editing
Figure 3 for FreeMask: Rethinking the Importance of Attention Masks for Zero-Shot Video Editing
Figure 4 for FreeMask: Rethinking the Importance of Attention Masks for Zero-Shot Video Editing
Viaarxiv icon

1-Bit FQT: Pushing the Limit of Fully Quantized Training to 1-bit

Add code
Aug 26, 2024
Figure 1 for 1-Bit FQT: Pushing the Limit of Fully Quantized Training to 1-bit
Figure 2 for 1-Bit FQT: Pushing the Limit of Fully Quantized Training to 1-bit
Figure 3 for 1-Bit FQT: Pushing the Limit of Fully Quantized Training to 1-bit
Figure 4 for 1-Bit FQT: Pushing the Limit of Fully Quantized Training to 1-bit
Viaarxiv icon

S^3D-NeRF: Single-Shot Speech-Driven Neural Radiance Field for High Fidelity Talking Head Synthesis

Add code
Aug 18, 2024
Figure 1 for S^3D-NeRF: Single-Shot Speech-Driven Neural Radiance Field for High Fidelity Talking Head Synthesis
Figure 2 for S^3D-NeRF: Single-Shot Speech-Driven Neural Radiance Field for High Fidelity Talking Head Synthesis
Figure 3 for S^3D-NeRF: Single-Shot Speech-Driven Neural Radiance Field for High Fidelity Talking Head Synthesis
Figure 4 for S^3D-NeRF: Single-Shot Speech-Driven Neural Radiance Field for High Fidelity Talking Head Synthesis
Viaarxiv icon

SUBLLM: A Novel Efficient Architecture with Token Sequence Subsampling for LLM

Add code
Jun 03, 2024
Viaarxiv icon

Accelerating Transformer Pre-Training with 2:4 Sparsity

Add code
Apr 02, 2024
Viaarxiv icon

Jetfire: Efficient and Accurate Transformer Pretraining with INT8 Data Flow and Per-Block Quantization

Add code
Mar 19, 2024
Viaarxiv icon