Picture for Akifumi Wachi

Akifumi Wachi

Inference-Aware Meta-Alignment of LLMs via Non-Linear GRPO

Add code
Feb 02, 2026
Viaarxiv icon

A Relative-Budget Theory for Reinforcement Learning with Verifiable Rewards in Large Language Model Reasoning

Add code
Feb 02, 2026
Viaarxiv icon

Cost-Minimized Label-Flipping Poisoning Attack to LLM Alignment

Add code
Nov 12, 2025
Viaarxiv icon

A Provable Approach for End-to-End Safe Reinforcement Learning

Add code
May 28, 2025
Viaarxiv icon

Offline Guarded Safe Reinforcement Learning for Medical Treatment Optimization Strategies

Add code
May 22, 2025
Viaarxiv icon

Target Return Optimizer for Multi-Game Decision Transformer

Add code
Mar 04, 2025
Viaarxiv icon

Vulnerability Mitigation for Safety-Aligned Language Models via Debiasing

Add code
Feb 04, 2025
Viaarxiv icon

Flipping-based Policy for Chance-Constrained Markov Decision Processes

Add code
Oct 09, 2024
Figure 1 for Flipping-based Policy for Chance-Constrained Markov Decision Processes
Figure 2 for Flipping-based Policy for Chance-Constrained Markov Decision Processes
Figure 3 for Flipping-based Policy for Chance-Constrained Markov Decision Processes
Figure 4 for Flipping-based Policy for Chance-Constrained Markov Decision Processes
Viaarxiv icon

Stepwise Alignment for Constrained Language Model Policy Optimization

Add code
Apr 17, 2024
Viaarxiv icon

A Survey of Constraint Formulations in Safe Reinforcement Learning

Add code
Feb 03, 2024
Viaarxiv icon