Picture for Sujian Li

Sujian Li

MPO: Boosting LLM Agents with Meta Plan Optimization

Add code
Mar 04, 2025
Viaarxiv icon

Chain-of-Thought Matters: Improving Long-Context Language Models with Reasoning Path Supervision

Add code
Feb 28, 2025
Viaarxiv icon

LongAttn: Selecting Long-context Training Data via Token-level Attention

Add code
Feb 24, 2025
Viaarxiv icon

More Tokens, Lower Precision: Towards the Optimal Token-Precision Trade-off in KV Cache Compression

Add code
Dec 17, 2024
Viaarxiv icon

VLRewardBench: A Challenging Benchmark for Vision-Language Generative Reward Models

Add code
Nov 26, 2024
Figure 1 for VLRewardBench: A Challenging Benchmark for Vision-Language Generative Reward Models
Figure 2 for VLRewardBench: A Challenging Benchmark for Vision-Language Generative Reward Models
Figure 3 for VLRewardBench: A Challenging Benchmark for Vision-Language Generative Reward Models
Figure 4 for VLRewardBench: A Challenging Benchmark for Vision-Language Generative Reward Models
Viaarxiv icon

AgentBank: Towards Generalized LLM Agents via Fine-Tuning on 50000+ Interaction Trajectories

Add code
Oct 10, 2024
Figure 1 for AgentBank: Towards Generalized LLM Agents via Fine-Tuning on 50000+ Interaction Trajectories
Figure 2 for AgentBank: Towards Generalized LLM Agents via Fine-Tuning on 50000+ Interaction Trajectories
Figure 3 for AgentBank: Towards Generalized LLM Agents via Fine-Tuning on 50000+ Interaction Trajectories
Figure 4 for AgentBank: Towards Generalized LLM Agents via Fine-Tuning on 50000+ Interaction Trajectories
Viaarxiv icon

Shapley Value-based Contrastive Alignment for Multimodal Information Extraction

Add code
Jul 25, 2024
Figure 1 for Shapley Value-based Contrastive Alignment for Multimodal Information Extraction
Figure 2 for Shapley Value-based Contrastive Alignment for Multimodal Information Extraction
Figure 3 for Shapley Value-based Contrastive Alignment for Multimodal Information Extraction
Figure 4 for Shapley Value-based Contrastive Alignment for Multimodal Information Extraction
Viaarxiv icon

The Good, The Bad, and The Greedy: Evaluation of LLMs Should Not Ignore Non-Determinism

Add code
Jul 15, 2024
Viaarxiv icon

EERPD: Leveraging Emotion and Emotion Regulation for Improving Personality Detection

Add code
Jun 23, 2024
Figure 1 for EERPD: Leveraging Emotion and Emotion Regulation for Improving Personality Detection
Figure 2 for EERPD: Leveraging Emotion and Emotion Regulation for Improving Personality Detection
Figure 3 for EERPD: Leveraging Emotion and Emotion Regulation for Improving Personality Detection
Figure 4 for EERPD: Leveraging Emotion and Emotion Regulation for Improving Personality Detection
Viaarxiv icon

Watch Every Step! LLM Agent Learning via Iterative Step-Level Process Refinement

Add code
Jun 17, 2024
Figure 1 for Watch Every Step! LLM Agent Learning via Iterative Step-Level Process Refinement
Figure 2 for Watch Every Step! LLM Agent Learning via Iterative Step-Level Process Refinement
Figure 3 for Watch Every Step! LLM Agent Learning via Iterative Step-Level Process Refinement
Figure 4 for Watch Every Step! LLM Agent Learning via Iterative Step-Level Process Refinement
Viaarxiv icon