Picture for Jie M. Zhang

Jie M. Zhang

University College London

Analyzing Message-Code Inconsistency in AI Coding Agent-Authored Pull Requests

Add code
Jan 08, 2026
Viaarxiv icon

Fairness Is Not Just Ethical: Performance Trade-Off via Data Correlation Tuning to Mitigate Bias in ML Software

Add code
Dec 19, 2025
Viaarxiv icon

Nexus: Execution-Grounded Multi-Agent Test Oracle Synthesis

Add code
Oct 30, 2025
Viaarxiv icon

Library Hallucinations in LLMs: Risk Analysis Grounded in Developer Queries

Add code
Sep 26, 2025
Figure 1 for Library Hallucinations in LLMs: Risk Analysis Grounded in Developer Queries
Figure 2 for Library Hallucinations in LLMs: Risk Analysis Grounded in Developer Queries
Figure 3 for Library Hallucinations in LLMs: Risk Analysis Grounded in Developer Queries
Figure 4 for Library Hallucinations in LLMs: Risk Analysis Grounded in Developer Queries
Viaarxiv icon

Large Language Models Miss the Multi-Agent Mark

Add code
May 27, 2025
Viaarxiv icon

AMQA: An Adversarial Dataset for Benchmarking Bias of LLMs in Medicine and Healthcare

Add code
May 26, 2025
Viaarxiv icon

EffiBench-X: A Multi-Language Benchmark for Measuring Efficiency of LLM-Generated Code

Add code
May 19, 2025
Figure 1 for EffiBench-X: A Multi-Language Benchmark for Measuring Efficiency of LLM-Generated Code
Figure 2 for EffiBench-X: A Multi-Language Benchmark for Measuring Efficiency of LLM-Generated Code
Figure 3 for EffiBench-X: A Multi-Language Benchmark for Measuring Efficiency of LLM-Generated Code
Figure 4 for EffiBench-X: A Multi-Language Benchmark for Measuring Efficiency of LLM-Generated Code
Viaarxiv icon

LLMs Love Python: A Study of LLMs' Bias for Programming Languages and Libraries

Add code
Mar 21, 2025
Figure 1 for LLMs Love Python: A Study of LLMs' Bias for Programming Languages and Libraries
Figure 2 for LLMs Love Python: A Study of LLMs' Bias for Programming Languages and Libraries
Figure 3 for LLMs Love Python: A Study of LLMs' Bias for Programming Languages and Libraries
Figure 4 for LLMs Love Python: A Study of LLMs' Bias for Programming Languages and Libraries
Viaarxiv icon

Hallucination Detection in Large Language Models with Metamorphic Relations

Add code
Feb 20, 2025
Figure 1 for Hallucination Detection in Large Language Models with Metamorphic Relations
Figure 2 for Hallucination Detection in Large Language Models with Metamorphic Relations
Figure 3 for Hallucination Detection in Large Language Models with Metamorphic Relations
Figure 4 for Hallucination Detection in Large Language Models with Metamorphic Relations
Viaarxiv icon

Fairness Aware Reinforcement Learning via Proximal Policy Optimization

Add code
Feb 06, 2025
Figure 1 for Fairness Aware Reinforcement Learning via Proximal Policy Optimization
Figure 2 for Fairness Aware Reinforcement Learning via Proximal Policy Optimization
Figure 3 for Fairness Aware Reinforcement Learning via Proximal Policy Optimization
Viaarxiv icon