Picture for Yangruibo Ding

Yangruibo Ding

OpenSage: Self-programming Agent Generation Engine

Add code
Feb 18, 2026
Viaarxiv icon

To Defend Against Cyber Attacks, We Must Teach AI Agents to Hack

Add code
Feb 01, 2026
Viaarxiv icon

SWE-Spot: Building Small Repo-Experts with Repository-Centric Learning

Add code
Jan 29, 2026
Viaarxiv icon

DevOps-Gym: Benchmarking AI Agents in Software DevOps Cycle

Add code
Jan 27, 2026
Viaarxiv icon

Co-PatcheR: Collaborative Software Patching with Component(s)-specific Small Reasoning Models

Add code
May 25, 2025
Figure 1 for Co-PatcheR: Collaborative Software Patching with Component(s)-specific Small Reasoning Models
Figure 2 for Co-PatcheR: Collaborative Software Patching with Component(s)-specific Small Reasoning Models
Figure 3 for Co-PatcheR: Collaborative Software Patching with Component(s)-specific Small Reasoning Models
Figure 4 for Co-PatcheR: Collaborative Software Patching with Component(s)-specific Small Reasoning Models
Viaarxiv icon

SemCoder: Training Code Language Models with Comprehensive Semantics

Add code
Jun 03, 2024
Viaarxiv icon

Vulnerability Detection with Code Language Models: How Far Are We?

Add code
Mar 27, 2024
Figure 1 for Vulnerability Detection with Code Language Models: How Far Are We?
Figure 2 for Vulnerability Detection with Code Language Models: How Far Are We?
Figure 3 for Vulnerability Detection with Code Language Models: How Far Are We?
Figure 4 for Vulnerability Detection with Code Language Models: How Far Are We?
Viaarxiv icon

CYCLE: Learning to Self-Refine the Code Generation

Add code
Mar 27, 2024
Viaarxiv icon

Beyond Accuracy: Evaluating Self-Consistency of Code Large Language Models with IdentityChain

Add code
Oct 21, 2023
Viaarxiv icon

CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion

Add code
Oct 17, 2023
Figure 1 for CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion
Figure 2 for CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion
Figure 3 for CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion
Figure 4 for CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion
Viaarxiv icon