Picture for Rui Yan

Rui Yan

More Expressive Attention with Negative Weights

Add code
Nov 14, 2024
Viaarxiv icon

Skywork-Reward: Bag of Tricks for Reward Modeling in LLMs

Add code
Oct 24, 2024
Viaarxiv icon

From a Tiny Slip to a Giant Leap: An LLM-Based Simulation for Fake News Evolution

Add code
Oct 24, 2024
Viaarxiv icon

Enhancing SNN-based Spatio-Temporal Learning: A Benchmark Dataset and Cross-Modality Attention Model

Add code
Oct 21, 2024
Viaarxiv icon

2D-TPE: Two-Dimensional Positional Encoding Enhances Table Understanding for Large Language Models

Add code
Sep 29, 2024
Viaarxiv icon

PEAR: Position-Embedding-Agnostic Attention Re-weighting Enhances Retrieval-Augmented Generation with Zero Inference Overhead

Add code
Sep 29, 2024
Figure 1 for PEAR: Position-Embedding-Agnostic Attention Re-weighting Enhances Retrieval-Augmented Generation with Zero Inference Overhead
Figure 2 for PEAR: Position-Embedding-Agnostic Attention Re-weighting Enhances Retrieval-Augmented Generation with Zero Inference Overhead
Figure 3 for PEAR: Position-Embedding-Agnostic Attention Re-weighting Enhances Retrieval-Augmented Generation with Zero Inference Overhead
Figure 4 for PEAR: Position-Embedding-Agnostic Attention Re-weighting Enhances Retrieval-Augmented Generation with Zero Inference Overhead
Viaarxiv icon

Language Models "Grok" to Copy

Add code
Sep 14, 2024
Viaarxiv icon

Locality-aware Cross-modal Correspondence Learning for Dense Audio-Visual Events Localization

Add code
Sep 12, 2024
Viaarxiv icon

Unlocking Decoding-time Controllability: Gradient-Free Multi-Objective Alignment with Contrastive Prompts

Add code
Aug 09, 2024
Viaarxiv icon

Towards Effective and Efficient Continual Pre-training of Large Language Models

Add code
Jul 26, 2024
Figure 1 for Towards Effective and Efficient Continual Pre-training of Large Language Models
Figure 2 for Towards Effective and Efficient Continual Pre-training of Large Language Models
Figure 3 for Towards Effective and Efficient Continual Pre-training of Large Language Models
Figure 4 for Towards Effective and Efficient Continual Pre-training of Large Language Models
Viaarxiv icon