Picture for Mohammad Sadegh Talebi

Mohammad Sadegh Talebi

No-regret Exploration in Shuffle Private Reinforcement Learning

Add code
Nov 18, 2024
Viaarxiv icon

Tractable Offline Learning of Regular Decision Processes

Add code
Sep 04, 2024
Viaarxiv icon

How to Shrink Confidence Sets for Many Equivalent Discrete Distributions?

Add code
Jul 22, 2024
Viaarxiv icon

Improved Exploration in Factored Average-Reward MDPs

Add code
Sep 09, 2020
Figure 1 for Improved Exploration in Factored Average-Reward MDPs
Figure 2 for Improved Exploration in Factored Average-Reward MDPs
Figure 3 for Improved Exploration in Factored Average-Reward MDPs
Viaarxiv icon

Tightening Exploration in Upper Confidence Reinforcement Learning

Add code
Apr 20, 2020
Figure 1 for Tightening Exploration in Upper Confidence Reinforcement Learning
Figure 2 for Tightening Exploration in Upper Confidence Reinforcement Learning
Figure 3 for Tightening Exploration in Upper Confidence Reinforcement Learning
Figure 4 for Tightening Exploration in Upper Confidence Reinforcement Learning
Viaarxiv icon

Model-Based Reinforcement Learning Exploiting State-Action Equivalence

Add code
Oct 09, 2019
Figure 1 for Model-Based Reinforcement Learning Exploiting State-Action Equivalence
Figure 2 for Model-Based Reinforcement Learning Exploiting State-Action Equivalence
Figure 3 for Model-Based Reinforcement Learning Exploiting State-Action Equivalence
Figure 4 for Model-Based Reinforcement Learning Exploiting State-Action Equivalence
Viaarxiv icon

Variance-Aware Regret Bounds for Undiscounted Reinforcement Learning in MDPs

Add code
Mar 05, 2018
Figure 1 for Variance-Aware Regret Bounds for Undiscounted Reinforcement Learning in MDPs
Figure 2 for Variance-Aware Regret Bounds for Undiscounted Reinforcement Learning in MDPs
Figure 3 for Variance-Aware Regret Bounds for Undiscounted Reinforcement Learning in MDPs
Figure 4 for Variance-Aware Regret Bounds for Undiscounted Reinforcement Learning in MDPs
Viaarxiv icon