Picture for Larkin Liu

Larkin Liu

Monte Carlo Planning for Stochastic Control on Constrained Markov Decision Processes

Add code
Jun 23, 2024
Viaarxiv icon

Approximate Nash Equilibrium Learning for n-Player Markov Games in Dynamic Pricing

Add code
Jul 13, 2022
Figure 1 for Approximate Nash Equilibrium Learning for n-Player Markov Games in Dynamic Pricing
Figure 2 for Approximate Nash Equilibrium Learning for n-Player Markov Games in Dynamic Pricing
Figure 3 for Approximate Nash Equilibrium Learning for n-Player Markov Games in Dynamic Pricing
Figure 4 for Approximate Nash Equilibrium Learning for n-Player Markov Games in Dynamic Pricing
Viaarxiv icon

An Extensible and Modular Design and Implementation of Monte Carlo Tree Search for the JVM

Add code
Jul 30, 2021
Figure 1 for An Extensible and Modular Design and Implementation of Monte Carlo Tree Search for the JVM
Figure 2 for An Extensible and Modular Design and Implementation of Monte Carlo Tree Search for the JVM
Figure 3 for An Extensible and Modular Design and Implementation of Monte Carlo Tree Search for the JVM
Figure 4 for An Extensible and Modular Design and Implementation of Monte Carlo Tree Search for the JVM
Viaarxiv icon

Improving the Performance of the LSTM and HMM Models via Hybridization

Add code
Jul 09, 2019
Figure 1 for Improving the Performance of the LSTM and HMM Models via Hybridization
Figure 2 for Improving the Performance of the LSTM and HMM Models via Hybridization
Figure 3 for Improving the Performance of the LSTM and HMM Models via Hybridization
Figure 4 for Improving the Performance of the LSTM and HMM Models via Hybridization
Viaarxiv icon

Multi-Armed Bandit Strategies for Non-Stationary Reward Distributions and Delayed Feedback Processes

Add code
Feb 22, 2019
Figure 1 for Multi-Armed Bandit Strategies for Non-Stationary Reward Distributions and Delayed Feedback Processes
Figure 2 for Multi-Armed Bandit Strategies for Non-Stationary Reward Distributions and Delayed Feedback Processes
Figure 3 for Multi-Armed Bandit Strategies for Non-Stationary Reward Distributions and Delayed Feedback Processes
Figure 4 for Multi-Armed Bandit Strategies for Non-Stationary Reward Distributions and Delayed Feedback Processes
Viaarxiv icon