Picture for Haishan Ye

Haishan Ye

ESSAM: A Novel Competitive Evolution Strategies Approach to Reinforcement Learning for Memory Efficient LLMs Fine-Tuning

Add code
Feb 01, 2026
Viaarxiv icon

Explicit and Non-asymptotic Query Complexities of Rank-Based Zeroth-order Algorithm on Stochastic Smooth Functions

Add code
Dec 22, 2025
Viaarxiv icon

Numerical Sensitivity and Robustness: Exploring the Flaws of Mathematical Reasoning in Large Language Models

Add code
Nov 11, 2025
Figure 1 for Numerical Sensitivity and Robustness: Exploring the Flaws of Mathematical Reasoning in Large Language Models
Figure 2 for Numerical Sensitivity and Robustness: Exploring the Flaws of Mathematical Reasoning in Large Language Models
Figure 3 for Numerical Sensitivity and Robustness: Exploring the Flaws of Mathematical Reasoning in Large Language Models
Figure 4 for Numerical Sensitivity and Robustness: Exploring the Flaws of Mathematical Reasoning in Large Language Models
Viaarxiv icon

MSCR: Exploring the Vulnerability of LLMs' Mathematical Reasoning Abilities Using Multi-Source Candidate Replacement

Add code
Nov 11, 2025
Viaarxiv icon

Frustratingly Easy Task-aware Pruning for Large Language Models

Add code
Oct 26, 2025
Figure 1 for Frustratingly Easy Task-aware Pruning for Large Language Models
Figure 2 for Frustratingly Easy Task-aware Pruning for Large Language Models
Figure 3 for Frustratingly Easy Task-aware Pruning for Large Language Models
Figure 4 for Frustratingly Easy Task-aware Pruning for Large Language Models
Viaarxiv icon

FZOO: Fast Zeroth-Order Optimizer for Fine-Tuning Large Language Models towards Adam-Scale Speed

Add code
Jun 10, 2025
Viaarxiv icon

Towards Understanding The Calibration Benefits of Sharpness-Aware Minimization

Add code
May 29, 2025
Viaarxiv icon

An Enhanced Zeroth-Order Stochastic Frank-Wolfe Framework for Constrained Finite-Sum Optimization

Add code
Jan 13, 2025
Viaarxiv icon

Double Variance Reduction: A Smoothing Trick for Composite Optimization Problems without First-Order Gradient

Add code
May 28, 2024
Figure 1 for Double Variance Reduction: A Smoothing Trick for Composite Optimization Problems without First-Order Gradient
Figure 2 for Double Variance Reduction: A Smoothing Trick for Composite Optimization Problems without First-Order Gradient
Figure 3 for Double Variance Reduction: A Smoothing Trick for Composite Optimization Problems without First-Order Gradient
Viaarxiv icon

Near-Optimal Distributed Minimax Optimization under the Second-Order Similarity

Add code
May 25, 2024
Viaarxiv icon