Picture for Kai Zhen

Kai Zhen

Indiana University Bloomington, Alexa Machine Learning, Amazon, USA

Saten: Sparse Augmented Tensor Networks for Post-Training Compression of Large Language Models

Add code
May 20, 2025
Viaarxiv icon

Wanda++: Pruning Large Language Models via Regional Gradients

Add code
Mar 06, 2025
Viaarxiv icon

QuZO: Quantized Zeroth-Order Fine-Tuning for Large Language Models

Add code
Feb 17, 2025
Viaarxiv icon

MaZO: Masked Zeroth-Order Optimization for Multi-Task Fine-Tuning of Large Language Models

Add code
Feb 17, 2025
Viaarxiv icon

AdaZeta: Adaptive Zeroth-Order Tensor-Train Adaption for Memory-Efficient Large Language Models Fine-Tuning

Add code
Jun 26, 2024
Figure 1 for AdaZeta: Adaptive Zeroth-Order Tensor-Train Adaption for Memory-Efficient Large Language Models Fine-Tuning
Figure 2 for AdaZeta: Adaptive Zeroth-Order Tensor-Train Adaption for Memory-Efficient Large Language Models Fine-Tuning
Figure 3 for AdaZeta: Adaptive Zeroth-Order Tensor-Train Adaption for Memory-Efficient Large Language Models Fine-Tuning
Figure 4 for AdaZeta: Adaptive Zeroth-Order Tensor-Train Adaption for Memory-Efficient Large Language Models Fine-Tuning
Viaarxiv icon

Sub-8-bit quantization for on-device speech recognition: a regularization-free approach

Add code
Oct 17, 2022
Figure 1 for Sub-8-bit quantization for on-device speech recognition: a regularization-free approach
Figure 2 for Sub-8-bit quantization for on-device speech recognition: a regularization-free approach
Figure 3 for Sub-8-bit quantization for on-device speech recognition: a regularization-free approach
Figure 4 for Sub-8-bit quantization for on-device speech recognition: a regularization-free approach
Viaarxiv icon

Sub-8-Bit Quantization Aware Training for 8-Bit Neural Network Accelerator with On-Device Speech Recognition

Add code
Jun 30, 2022
Figure 1 for Sub-8-Bit Quantization Aware Training for 8-Bit Neural Network Accelerator with On-Device Speech Recognition
Figure 2 for Sub-8-Bit Quantization Aware Training for 8-Bit Neural Network Accelerator with On-Device Speech Recognition
Figure 3 for Sub-8-Bit Quantization Aware Training for 8-Bit Neural Network Accelerator with On-Device Speech Recognition
Figure 4 for Sub-8-Bit Quantization Aware Training for 8-Bit Neural Network Accelerator with On-Device Speech Recognition
Viaarxiv icon

Scalable and Efficient Neural Speech Coding

Add code
Mar 27, 2021
Figure 1 for Scalable and Efficient Neural Speech Coding
Figure 2 for Scalable and Efficient Neural Speech Coding
Figure 3 for Scalable and Efficient Neural Speech Coding
Figure 4 for Scalable and Efficient Neural Speech Coding
Viaarxiv icon

Sparsification via Compressed Sensing for Automatic Speech Recognition

Add code
Feb 09, 2021
Figure 1 for Sparsification via Compressed Sensing for Automatic Speech Recognition
Figure 2 for Sparsification via Compressed Sensing for Automatic Speech Recognition
Figure 3 for Sparsification via Compressed Sensing for Automatic Speech Recognition
Figure 4 for Sparsification via Compressed Sensing for Automatic Speech Recognition
Viaarxiv icon

Psychoacoustic Calibration of Loss Functions for Efficient End-to-End Neural Audio Coding

Add code
Dec 31, 2020
Figure 1 for Psychoacoustic Calibration of Loss Functions for Efficient End-to-End Neural Audio Coding
Figure 2 for Psychoacoustic Calibration of Loss Functions for Efficient End-to-End Neural Audio Coding
Figure 3 for Psychoacoustic Calibration of Loss Functions for Efficient End-to-End Neural Audio Coding
Figure 4 for Psychoacoustic Calibration of Loss Functions for Efficient End-to-End Neural Audio Coding
Viaarxiv icon