Picture for Qingru Zhang

Qingru Zhang

Model Tells Itself Where to Attend: Faithfulness Meets Automatic Attention Steering

Add code
Sep 16, 2024
Figure 1 for Model Tells Itself Where to Attend: Faithfulness Meets Automatic Attention Steering
Figure 2 for Model Tells Itself Where to Attend: Faithfulness Meets Automatic Attention Steering
Figure 3 for Model Tells Itself Where to Attend: Faithfulness Meets Automatic Attention Steering
Figure 4 for Model Tells Itself Where to Attend: Faithfulness Meets Automatic Attention Steering
Viaarxiv icon

Robust Reinforcement Learning from Corrupted Human Feedback

Add code
Jun 21, 2024
Viaarxiv icon

GEAR: An Efficient KV Cache Compression Recipe for Near-Lossless Generative Inference of LLM

Add code
Mar 11, 2024
Viaarxiv icon

Tell Your Model Where to Attend: Post-hoc Attention Steering for LLMs

Add code
Nov 03, 2023
Viaarxiv icon

Efficient Long-Range Transformers: You Need to Attend More, but Not Necessarily at Every Layer

Add code
Oct 19, 2023
Viaarxiv icon

Robust Multi-Agent Reinforcement Learning via Adversarial Regularization: Theoretical Foundation and Stable Algorithms

Add code
Oct 16, 2023
Viaarxiv icon

LoSparse: Structured Compression of Large Language Models based on Low-Rank and Sparse Approximation

Add code
Jun 26, 2023
Viaarxiv icon

Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning

Add code
Mar 18, 2023
Viaarxiv icon

Less is More: Task-aware Layer-wise Distillation for Language Model Compression

Add code
Oct 05, 2022
Figure 1 for Less is More: Task-aware Layer-wise Distillation for Language Model Compression
Figure 2 for Less is More: Task-aware Layer-wise Distillation for Language Model Compression
Figure 3 for Less is More: Task-aware Layer-wise Distillation for Language Model Compression
Figure 4 for Less is More: Task-aware Layer-wise Distillation for Language Model Compression
Viaarxiv icon

PLATON: Pruning Large Transformer Models with Upper Confidence Bound of Weight Importance

Add code
Jun 25, 2022
Figure 1 for PLATON: Pruning Large Transformer Models with Upper Confidence Bound of Weight Importance
Figure 2 for PLATON: Pruning Large Transformer Models with Upper Confidence Bound of Weight Importance
Figure 3 for PLATON: Pruning Large Transformer Models with Upper Confidence Bound of Weight Importance
Figure 4 for PLATON: Pruning Large Transformer Models with Upper Confidence Bound of Weight Importance
Viaarxiv icon