Picture for Linping Qu

Linping Qu

KVTuner: Sensitivity-Aware Layer-wise Mixed Precision KV Cache Quantization for Efficient and Nearly Lossless LLM Inference

Add code
Feb 06, 2025
Figure 1 for KVTuner: Sensitivity-Aware Layer-wise Mixed Precision KV Cache Quantization for Efficient and Nearly Lossless LLM Inference
Figure 2 for KVTuner: Sensitivity-Aware Layer-wise Mixed Precision KV Cache Quantization for Efficient and Nearly Lossless LLM Inference
Figure 3 for KVTuner: Sensitivity-Aware Layer-wise Mixed Precision KV Cache Quantization for Efficient and Nearly Lossless LLM Inference
Figure 4 for KVTuner: Sensitivity-Aware Layer-wise Mixed Precision KV Cache Quantization for Efficient and Nearly Lossless LLM Inference
Viaarxiv icon

FedAQ: Communication-Efficient Federated Edge Learning via Joint Uplink and Downlink Adaptive Quantization

Add code
Jun 26, 2024
Figure 1 for FedAQ: Communication-Efficient Federated Edge Learning via Joint Uplink and Downlink Adaptive Quantization
Figure 2 for FedAQ: Communication-Efficient Federated Edge Learning via Joint Uplink and Downlink Adaptive Quantization
Figure 3 for FedAQ: Communication-Efficient Federated Edge Learning via Joint Uplink and Downlink Adaptive Quantization
Figure 4 for FedAQ: Communication-Efficient Federated Edge Learning via Joint Uplink and Downlink Adaptive Quantization
Viaarxiv icon

How Robust is Federated Learning to Communication Error? A Comparison Study Between Uplink and Downlink Channels

Add code
Oct 25, 2023
Viaarxiv icon

FedDQ: Communication-Efficient Federated Learning with Descending Quantization

Add code
Oct 13, 2021
Figure 1 for FedDQ: Communication-Efficient Federated Learning with Descending Quantization
Figure 2 for FedDQ: Communication-Efficient Federated Learning with Descending Quantization
Figure 3 for FedDQ: Communication-Efficient Federated Learning with Descending Quantization
Figure 4 for FedDQ: Communication-Efficient Federated Learning with Descending Quantization
Viaarxiv icon