Picture for Ngai Wong

Ngai Wong

ParallelComp: Parallel Long-Context Compressor for Length Extrapolation

Add code
Feb 20, 2025
Viaarxiv icon

QuZO: Quantized Zeroth-Order Fine-Tuning for Large Language Models

Add code
Feb 17, 2025
Viaarxiv icon

Quantization Meets Reasoning: Exploring LLM Low-Bit Quantization Degradation for Mathematical Reasoning

Add code
Jan 06, 2025
Viaarxiv icon

LLM-Neo: Parameter Efficient Knowledge Distillation for Large Language Models

Add code
Nov 11, 2024
Viaarxiv icon

Autoregressive Models in Vision: A Survey

Add code
Nov 08, 2024
Figure 1 for Autoregressive Models in Vision: A Survey
Figure 2 for Autoregressive Models in Vision: A Survey
Figure 3 for Autoregressive Models in Vision: A Survey
Figure 4 for Autoregressive Models in Vision: A Survey
Viaarxiv icon

MCUBERT: Memory-Efficient BERT Inference on Commodity Microcontrollers

Add code
Oct 23, 2024
Figure 1 for MCUBERT: Memory-Efficient BERT Inference on Commodity Microcontrollers
Figure 2 for MCUBERT: Memory-Efficient BERT Inference on Commodity Microcontrollers
Figure 3 for MCUBERT: Memory-Efficient BERT Inference on Commodity Microcontrollers
Figure 4 for MCUBERT: Memory-Efficient BERT Inference on Commodity Microcontrollers
Viaarxiv icon

UNComp: Uncertainty-Aware Long-Context Compressor for Efficient Large Language Model Inference

Add code
Oct 04, 2024
Figure 1 for UNComp: Uncertainty-Aware Long-Context Compressor for Efficient Large Language Model Inference
Figure 2 for UNComp: Uncertainty-Aware Long-Context Compressor for Efficient Large Language Model Inference
Figure 3 for UNComp: Uncertainty-Aware Long-Context Compressor for Efficient Large Language Model Inference
Figure 4 for UNComp: Uncertainty-Aware Long-Context Compressor for Efficient Large Language Model Inference
Viaarxiv icon

UncertaintyRAG: Span-Level Uncertainty Enhanced Long-Context Modeling for Retrieval-Augmented Generation

Add code
Oct 03, 2024
Figure 1 for UncertaintyRAG: Span-Level Uncertainty Enhanced Long-Context Modeling for Retrieval-Augmented Generation
Figure 2 for UncertaintyRAG: Span-Level Uncertainty Enhanced Long-Context Modeling for Retrieval-Augmented Generation
Figure 3 for UncertaintyRAG: Span-Level Uncertainty Enhanced Long-Context Modeling for Retrieval-Augmented Generation
Figure 4 for UncertaintyRAG: Span-Level Uncertainty Enhanced Long-Context Modeling for Retrieval-Augmented Generation
Viaarxiv icon

A Survey on the Honesty of Large Language Models

Add code
Sep 27, 2024
Figure 1 for A Survey on the Honesty of Large Language Models
Figure 2 for A Survey on the Honesty of Large Language Models
Figure 3 for A Survey on the Honesty of Large Language Models
Figure 4 for A Survey on the Honesty of Large Language Models
Viaarxiv icon

LLM-Barber: Block-Aware Rebuilder for Sparsity Mask in One-Shot for Large Language Models

Add code
Aug 20, 2024
Viaarxiv icon