Picture for Yongbin Li

Yongbin Li

DEMO: Reframing Dialogue Interaction with Fine-grained Element Modeling

Add code
Dec 06, 2024
Viaarxiv icon

LLMs as Continuous Learners: Improving the Reproduction of Defective Code in Software Issues

Add code
Nov 21, 2024
Viaarxiv icon

IOPO: Empowering LLMs with Complex Instruction Following via Input-Output Preference Optimization

Add code
Nov 09, 2024
Figure 1 for IOPO: Empowering LLMs with Complex Instruction Following via Input-Output Preference Optimization
Figure 2 for IOPO: Empowering LLMs with Complex Instruction Following via Input-Output Preference Optimization
Figure 3 for IOPO: Empowering LLMs with Complex Instruction Following via Input-Output Preference Optimization
Figure 4 for IOPO: Empowering LLMs with Complex Instruction Following via Input-Output Preference Optimization
Viaarxiv icon

Lingma SWE-GPT: An Open Development-Process-Centric Language Model for Automated Software Improvement

Add code
Nov 01, 2024
Figure 1 for Lingma SWE-GPT: An Open Development-Process-Centric Language Model for Automated Software Improvement
Figure 2 for Lingma SWE-GPT: An Open Development-Process-Centric Language Model for Automated Software Improvement
Figure 3 for Lingma SWE-GPT: An Open Development-Process-Centric Language Model for Automated Software Improvement
Figure 4 for Lingma SWE-GPT: An Open Development-Process-Centric Language Model for Automated Software Improvement
Viaarxiv icon

Transferable Post-training via Inverse Value Learning

Add code
Oct 28, 2024
Figure 1 for Transferable Post-training via Inverse Value Learning
Figure 2 for Transferable Post-training via Inverse Value Learning
Figure 3 for Transferable Post-training via Inverse Value Learning
Figure 4 for Transferable Post-training via Inverse Value Learning
Viaarxiv icon

On the Role of Attention Heads in Large Language Model Safety

Add code
Oct 17, 2024
Figure 1 for On the Role of Attention Heads in Large Language Model Safety
Figure 2 for On the Role of Attention Heads in Large Language Model Safety
Figure 3 for On the Role of Attention Heads in Large Language Model Safety
Figure 4 for On the Role of Attention Heads in Large Language Model Safety
Viaarxiv icon

StructRAG: Boosting Knowledge Intensive Reasoning of LLMs via Inference-time Hybrid Information Structurization

Add code
Oct 11, 2024
Figure 1 for StructRAG: Boosting Knowledge Intensive Reasoning of LLMs via Inference-time Hybrid Information Structurization
Figure 2 for StructRAG: Boosting Knowledge Intensive Reasoning of LLMs via Inference-time Hybrid Information Structurization
Figure 3 for StructRAG: Boosting Knowledge Intensive Reasoning of LLMs via Inference-time Hybrid Information Structurization
Figure 4 for StructRAG: Boosting Knowledge Intensive Reasoning of LLMs via Inference-time Hybrid Information Structurization
Viaarxiv icon

In-Context Transfer Learning: Demonstration Synthesis by Transferring Similar Tasks

Add code
Oct 02, 2024
Figure 1 for In-Context Transfer Learning: Demonstration Synthesis by Transferring Similar Tasks
Figure 2 for In-Context Transfer Learning: Demonstration Synthesis by Transferring Similar Tasks
Figure 3 for In-Context Transfer Learning: Demonstration Synthesis by Transferring Similar Tasks
Figure 4 for In-Context Transfer Learning: Demonstration Synthesis by Transferring Similar Tasks
Viaarxiv icon

Codev-Bench: How Do LLMs Understand Developer-Centric Code Completion?

Add code
Oct 02, 2024
Figure 1 for Codev-Bench: How Do LLMs Understand Developer-Centric Code Completion?
Figure 2 for Codev-Bench: How Do LLMs Understand Developer-Centric Code Completion?
Figure 3 for Codev-Bench: How Do LLMs Understand Developer-Centric Code Completion?
Figure 4 for Codev-Bench: How Do LLMs Understand Developer-Centric Code Completion?
Viaarxiv icon

Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining

Add code
Oct 01, 2024
Figure 1 for Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining
Figure 2 for Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining
Figure 3 for Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining
Figure 4 for Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining
Viaarxiv icon