Picture for Toshinori Kitamura

Toshinori Kitamura

Near-Optimal Policy Identification in Robust Constrained Markov Decision Processes via Epigraph Form

Add code
Sep 02, 2024
Viaarxiv icon

A Policy Gradient Primal-Dual Algorithm for Constrained MDPs with Uniform PAC Guarantees

Add code
Feb 02, 2024
Viaarxiv icon

Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice

Add code
May 22, 2023
Viaarxiv icon

KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal

Add code
May 27, 2022
Figure 1 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 2 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 3 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 4 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Viaarxiv icon

ShinRL: A Library for Evaluating RL Algorithms from Theoretical and Practical Perspectives

Add code
Dec 08, 2021
Figure 1 for ShinRL: A Library for Evaluating RL Algorithms from Theoretical and Practical Perspectives
Figure 2 for ShinRL: A Library for Evaluating RL Algorithms from Theoretical and Practical Perspectives
Figure 3 for ShinRL: A Library for Evaluating RL Algorithms from Theoretical and Practical Perspectives
Figure 4 for ShinRL: A Library for Evaluating RL Algorithms from Theoretical and Practical Perspectives
Viaarxiv icon

Geometric Value Iteration: Dynamic Error-Aware KL Regularization for Reinforcement Learning

Add code
Jul 16, 2021
Figure 1 for Geometric Value Iteration: Dynamic Error-Aware KL Regularization for Reinforcement Learning
Figure 2 for Geometric Value Iteration: Dynamic Error-Aware KL Regularization for Reinforcement Learning
Figure 3 for Geometric Value Iteration: Dynamic Error-Aware KL Regularization for Reinforcement Learning
Figure 4 for Geometric Value Iteration: Dynamic Error-Aware KL Regularization for Reinforcement Learning
Viaarxiv icon

Cautious Policy Programming: Exploiting KL Regularization in Monotonic Policy Improvement for Reinforcement Learning

Add code
Jul 13, 2021
Figure 1 for Cautious Policy Programming: Exploiting KL Regularization in Monotonic Policy Improvement for Reinforcement Learning
Figure 2 for Cautious Policy Programming: Exploiting KL Regularization in Monotonic Policy Improvement for Reinforcement Learning
Figure 3 for Cautious Policy Programming: Exploiting KL Regularization in Monotonic Policy Improvement for Reinforcement Learning
Figure 4 for Cautious Policy Programming: Exploiting KL Regularization in Monotonic Policy Improvement for Reinforcement Learning
Viaarxiv icon

Cautious Actor-Critic

Add code
Jul 12, 2021
Figure 1 for Cautious Actor-Critic
Figure 2 for Cautious Actor-Critic
Figure 3 for Cautious Actor-Critic
Figure 4 for Cautious Actor-Critic
Viaarxiv icon