Picture for Xiuying Wei

Xiuying Wei

Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis

Add code
Jul 13, 2024
Figure 1 for Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis
Figure 2 for Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis
Figure 3 for Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis
Figure 4 for Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis
Viaarxiv icon

Building on Efficient Foundations: Effectively Training LLMs with Structured Feedforward Layers

Add code
Jun 24, 2024
Viaarxiv icon

Selective Focus: Investigating Semantics Sensitivity in Post-training Quantization for Lane Detection

Add code
May 10, 2024
Viaarxiv icon

Fast and Controllable Post-training Sparsity: Learning Optimal Sparsity Allocation with Global Constraint in Minutes

Add code
May 09, 2024
Viaarxiv icon

QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language Models

Add code
Oct 12, 2023
Viaarxiv icon

Lossy and Lossless Post-training Model Size Compression

Add code
Aug 08, 2023
Viaarxiv icon

Outlier Suppression+: Accurate quantization of large language models by equivalent and optimal shifting and scaling

Add code
Apr 18, 2023
Viaarxiv icon

Outlier Suppression: Pushing the Limit of Low-bit Transformer Language Models

Add code
Sep 27, 2022
Figure 1 for Outlier Suppression: Pushing the Limit of Low-bit Transformer Language Models
Figure 2 for Outlier Suppression: Pushing the Limit of Low-bit Transformer Language Models
Figure 3 for Outlier Suppression: Pushing the Limit of Low-bit Transformer Language Models
Figure 4 for Outlier Suppression: Pushing the Limit of Low-bit Transformer Language Models
Viaarxiv icon

QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization

Add code
Mar 11, 2022
Figure 1 for QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization
Figure 2 for QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization
Figure 3 for QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization
Figure 4 for QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization
Viaarxiv icon