Picture for Brahma S. Pavse

Brahma S. Pavse

Stable Offline Value Function Learning with Bisimulation-based Representations

Add code
Oct 02, 2024
Figure 1 for Stable Offline Value Function Learning with Bisimulation-based Representations
Figure 2 for Stable Offline Value Function Learning with Bisimulation-based Representations
Figure 3 for Stable Offline Value Function Learning with Bisimulation-based Representations
Figure 4 for Stable Offline Value Function Learning with Bisimulation-based Representations
Viaarxiv icon

State-Action Similarity-Based Representations for Off-Policy Evaluation

Add code
Oct 27, 2023
Figure 1 for State-Action Similarity-Based Representations for Off-Policy Evaluation
Figure 2 for State-Action Similarity-Based Representations for Off-Policy Evaluation
Figure 3 for State-Action Similarity-Based Representations for Off-Policy Evaluation
Figure 4 for State-Action Similarity-Based Representations for Off-Policy Evaluation
Viaarxiv icon

Tackling Unbounded State Spaces in Continuing Task Reinforcement Learning

Add code
Jun 02, 2023
Figure 1 for Tackling Unbounded State Spaces in Continuing Task Reinforcement Learning
Figure 2 for Tackling Unbounded State Spaces in Continuing Task Reinforcement Learning
Figure 3 for Tackling Unbounded State Spaces in Continuing Task Reinforcement Learning
Figure 4 for Tackling Unbounded State Spaces in Continuing Task Reinforcement Learning
Viaarxiv icon

Scaling Marginalized Importance Sampling to High-Dimensional State-Spaces via State Abstraction

Add code
Dec 14, 2022
Figure 1 for Scaling Marginalized Importance Sampling to High-Dimensional State-Spaces via State Abstraction
Figure 2 for Scaling Marginalized Importance Sampling to High-Dimensional State-Spaces via State Abstraction
Figure 3 for Scaling Marginalized Importance Sampling to High-Dimensional State-Spaces via State Abstraction
Figure 4 for Scaling Marginalized Importance Sampling to High-Dimensional State-Spaces via State Abstraction
Viaarxiv icon

RIDM: Reinforced Inverse Dynamics Modeling for Learning from a Single Observed Demonstration

Add code
Jul 01, 2019
Figure 1 for RIDM: Reinforced Inverse Dynamics Modeling for Learning from a Single Observed Demonstration
Figure 2 for RIDM: Reinforced Inverse Dynamics Modeling for Learning from a Single Observed Demonstration
Figure 3 for RIDM: Reinforced Inverse Dynamics Modeling for Learning from a Single Observed Demonstration
Figure 4 for RIDM: Reinforced Inverse Dynamics Modeling for Learning from a Single Observed Demonstration
Viaarxiv icon