Picture for Ngai Wong

Ngai Wong

DnLUT: Ultra-Efficient Color Image Denoising via Channel-Aware Lookup Tables

Add code
Mar 20, 2025
Viaarxiv icon

HaLoRA: Hardware-aware Low-Rank Adaptation for Large Language Models Based on Hybrid Compute-in-Memory Architecture

Add code
Feb 27, 2025
Viaarxiv icon

ParallelComp: Parallel Long-Context Compressor for Length Extrapolation

Add code
Feb 20, 2025
Viaarxiv icon

QuZO: Quantized Zeroth-Order Fine-Tuning for Large Language Models

Add code
Feb 17, 2025
Viaarxiv icon

Quantization Meets Reasoning: Exploring LLM Low-Bit Quantization Degradation for Mathematical Reasoning

Add code
Jan 06, 2025
Viaarxiv icon

LLM-Neo: Parameter Efficient Knowledge Distillation for Large Language Models

Add code
Nov 11, 2024
Viaarxiv icon

Autoregressive Models in Vision: A Survey

Add code
Nov 08, 2024
Figure 1 for Autoregressive Models in Vision: A Survey
Figure 2 for Autoregressive Models in Vision: A Survey
Figure 3 for Autoregressive Models in Vision: A Survey
Figure 4 for Autoregressive Models in Vision: A Survey
Viaarxiv icon

MCUBERT: Memory-Efficient BERT Inference on Commodity Microcontrollers

Add code
Oct 23, 2024
Figure 1 for MCUBERT: Memory-Efficient BERT Inference on Commodity Microcontrollers
Figure 2 for MCUBERT: Memory-Efficient BERT Inference on Commodity Microcontrollers
Figure 3 for MCUBERT: Memory-Efficient BERT Inference on Commodity Microcontrollers
Figure 4 for MCUBERT: Memory-Efficient BERT Inference on Commodity Microcontrollers
Viaarxiv icon

UNComp: Uncertainty-Aware Long-Context Compressor for Efficient Large Language Model Inference

Add code
Oct 04, 2024
Figure 1 for UNComp: Uncertainty-Aware Long-Context Compressor for Efficient Large Language Model Inference
Figure 2 for UNComp: Uncertainty-Aware Long-Context Compressor for Efficient Large Language Model Inference
Figure 3 for UNComp: Uncertainty-Aware Long-Context Compressor for Efficient Large Language Model Inference
Figure 4 for UNComp: Uncertainty-Aware Long-Context Compressor for Efficient Large Language Model Inference
Viaarxiv icon

UncertaintyRAG: Span-Level Uncertainty Enhanced Long-Context Modeling for Retrieval-Augmented Generation

Add code
Oct 03, 2024
Figure 1 for UncertaintyRAG: Span-Level Uncertainty Enhanced Long-Context Modeling for Retrieval-Augmented Generation
Figure 2 for UncertaintyRAG: Span-Level Uncertainty Enhanced Long-Context Modeling for Retrieval-Augmented Generation
Figure 3 for UncertaintyRAG: Span-Level Uncertainty Enhanced Long-Context Modeling for Retrieval-Augmented Generation
Figure 4 for UncertaintyRAG: Span-Level Uncertainty Enhanced Long-Context Modeling for Retrieval-Augmented Generation
Viaarxiv icon