Picture for Matteo Papini

Matteo Papini

How Log-Barrier Helps Exploration in Policy Optimization

Add code
Mar 16, 2026
Viaarxiv icon

Impact of Connectivity on Laplacian Representations in Reinforcement Learning

Add code
Mar 09, 2026
Viaarxiv icon

Do It for HER: First-Order Temporal Logic Reward Specification in Reinforcement Learning (Extended Version)

Add code
Feb 05, 2026
Viaarxiv icon

Reusing Trajectories in Policy Gradients Enables Fast Convergence

Add code
Jun 06, 2025
Figure 1 for Reusing Trajectories in Policy Gradients Enables Fast Convergence
Figure 2 for Reusing Trajectories in Policy Gradients Enables Fast Convergence
Figure 3 for Reusing Trajectories in Policy Gradients Enables Fast Convergence
Figure 4 for Reusing Trajectories in Policy Gradients Enables Fast Convergence
Viaarxiv icon

Learning Deterministic Policies with Policy Gradients in Constrained Markov Decision Processes

Add code
Jun 06, 2025
Figure 1 for Learning Deterministic Policies with Policy Gradients in Constrained Markov Decision Processes
Figure 2 for Learning Deterministic Policies with Policy Gradients in Constrained Markov Decision Processes
Figure 3 for Learning Deterministic Policies with Policy Gradients in Constrained Markov Decision Processes
Figure 4 for Learning Deterministic Policies with Policy Gradients in Constrained Markov Decision Processes
Viaarxiv icon

Statistical Analysis of Policy Space Compression Problem

Add code
Nov 15, 2024
Viaarxiv icon

Local Linearity: the Key for No-regret Reinforcement Learning in Continuous MDPs

Add code
Oct 31, 2024
Viaarxiv icon

Last-Iterate Global Convergence of Policy Gradients for Constrained Reinforcement Learning

Add code
Jul 15, 2024
Viaarxiv icon

Projection by Convolution: Optimal Sample Complexity for Reinforcement Learning in Continuous-Space MDPs

Add code
May 10, 2024
Viaarxiv icon

Policy Gradient with Active Importance Sampling

Add code
May 09, 2024
Figure 1 for Policy Gradient with Active Importance Sampling
Figure 2 for Policy Gradient with Active Importance Sampling
Figure 3 for Policy Gradient with Active Importance Sampling
Figure 4 for Policy Gradient with Active Importance Sampling
Viaarxiv icon