Picture for Yongbin Li

Yongbin Li

Lingma SWE-GPT: An Open Development-Process-Centric Language Model for Automated Software Improvement

Add code
Nov 01, 2024
Viaarxiv icon

Transferable Post-training via Inverse Value Learning

Add code
Oct 28, 2024
Figure 1 for Transferable Post-training via Inverse Value Learning
Figure 2 for Transferable Post-training via Inverse Value Learning
Figure 3 for Transferable Post-training via Inverse Value Learning
Figure 4 for Transferable Post-training via Inverse Value Learning
Viaarxiv icon

On the Role of Attention Heads in Large Language Model Safety

Add code
Oct 17, 2024
Figure 1 for On the Role of Attention Heads in Large Language Model Safety
Figure 2 for On the Role of Attention Heads in Large Language Model Safety
Figure 3 for On the Role of Attention Heads in Large Language Model Safety
Figure 4 for On the Role of Attention Heads in Large Language Model Safety
Viaarxiv icon

StructRAG: Boosting Knowledge Intensive Reasoning of LLMs via Inference-time Hybrid Information Structurization

Add code
Oct 11, 2024
Figure 1 for StructRAG: Boosting Knowledge Intensive Reasoning of LLMs via Inference-time Hybrid Information Structurization
Figure 2 for StructRAG: Boosting Knowledge Intensive Reasoning of LLMs via Inference-time Hybrid Information Structurization
Figure 3 for StructRAG: Boosting Knowledge Intensive Reasoning of LLMs via Inference-time Hybrid Information Structurization
Figure 4 for StructRAG: Boosting Knowledge Intensive Reasoning of LLMs via Inference-time Hybrid Information Structurization
Viaarxiv icon

Codev-Bench: How Do LLMs Understand Developer-Centric Code Completion?

Add code
Oct 02, 2024
Figure 1 for Codev-Bench: How Do LLMs Understand Developer-Centric Code Completion?
Figure 2 for Codev-Bench: How Do LLMs Understand Developer-Centric Code Completion?
Figure 3 for Codev-Bench: How Do LLMs Understand Developer-Centric Code Completion?
Figure 4 for Codev-Bench: How Do LLMs Understand Developer-Centric Code Completion?
Viaarxiv icon

In-Context Transfer Learning: Demonstration Synthesis by Transferring Similar Tasks

Add code
Oct 02, 2024
Viaarxiv icon

Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining

Add code
Oct 01, 2024
Figure 1 for Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining
Figure 2 for Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining
Figure 3 for Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining
Figure 4 for Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining
Viaarxiv icon

The Imperative of Conversation Analysis in the Era of LLMs: A Survey of Tasks, Techniques, and Trends

Add code
Sep 21, 2024
Figure 1 for The Imperative of Conversation Analysis in the Era of LLMs: A Survey of Tasks, Techniques, and Trends
Figure 2 for The Imperative of Conversation Analysis in the Era of LLMs: A Survey of Tasks, Techniques, and Trends
Figure 3 for The Imperative of Conversation Analysis in the Era of LLMs: A Survey of Tasks, Techniques, and Trends
Figure 4 for The Imperative of Conversation Analysis in the Era of LLMs: A Survey of Tasks, Techniques, and Trends
Viaarxiv icon

MMEvol: Empowering Multimodal Large Language Models with Evol-Instruct

Add code
Sep 09, 2024
Figure 1 for MMEvol: Empowering Multimodal Large Language Models with Evol-Instruct
Figure 2 for MMEvol: Empowering Multimodal Large Language Models with Evol-Instruct
Figure 3 for MMEvol: Empowering Multimodal Large Language Models with Evol-Instruct
Figure 4 for MMEvol: Empowering Multimodal Large Language Models with Evol-Instruct
Viaarxiv icon

Extend Model Merging from Fine-Tuned to Pre-Trained Large Language Models via Weight Disentanglement

Add code
Aug 06, 2024
Figure 1 for Extend Model Merging from Fine-Tuned to Pre-Trained Large Language Models via Weight Disentanglement
Figure 2 for Extend Model Merging from Fine-Tuned to Pre-Trained Large Language Models via Weight Disentanglement
Figure 3 for Extend Model Merging from Fine-Tuned to Pre-Trained Large Language Models via Weight Disentanglement
Figure 4 for Extend Model Merging from Fine-Tuned to Pre-Trained Large Language Models via Weight Disentanglement
Viaarxiv icon