Picture for Yusuke Narita

Yusuke Narita

Off-Policy Evaluation of Ranking Policies under Diverse User Behavior

Add code
Jun 26, 2023
Viaarxiv icon

Counterfactual Learning with General Data-generating Policies

Add code
Dec 04, 2022
Viaarxiv icon

Policy-Adaptive Estimator Selection for Off-Policy Evaluation

Add code
Nov 25, 2022
Viaarxiv icon

Doubly Robust Off-Policy Evaluation for Ranking Policies under the Cascade Behavior Model

Add code
Feb 03, 2022
Figure 1 for Doubly Robust Off-Policy Evaluation for Ranking Policies under the Cascade Behavior Model
Figure 2 for Doubly Robust Off-Policy Evaluation for Ranking Policies under the Cascade Behavior Model
Figure 3 for Doubly Robust Off-Policy Evaluation for Ranking Policies under the Cascade Behavior Model
Figure 4 for Doubly Robust Off-Policy Evaluation for Ranking Policies under the Cascade Behavior Model
Viaarxiv icon

Evaluating the Robustness of Off-Policy Evaluation

Add code
Aug 31, 2021
Figure 1 for Evaluating the Robustness of Off-Policy Evaluation
Figure 2 for Evaluating the Robustness of Off-Policy Evaluation
Figure 3 for Evaluating the Robustness of Off-Policy Evaluation
Figure 4 for Evaluating the Robustness of Off-Policy Evaluation
Viaarxiv icon

Algorithm is Experiment: Machine Learning, Market Design, and Policy Eligibility Rules

Add code
Apr 26, 2021
Figure 1 for Algorithm is Experiment: Machine Learning, Market Design, and Policy Eligibility Rules
Figure 2 for Algorithm is Experiment: Machine Learning, Market Design, and Policy Eligibility Rules
Figure 3 for Algorithm is Experiment: Machine Learning, Market Design, and Policy Eligibility Rules
Figure 4 for Algorithm is Experiment: Machine Learning, Market Design, and Policy Eligibility Rules
Viaarxiv icon

A Large-scale Open Dataset for Bandit Algorithms

Add code
Aug 17, 2020
Figure 1 for A Large-scale Open Dataset for Bandit Algorithms
Figure 2 for A Large-scale Open Dataset for Bandit Algorithms
Viaarxiv icon

Safe Counterfactual Reinforcement Learning

Add code
Feb 20, 2020
Figure 1 for Safe Counterfactual Reinforcement Learning
Figure 2 for Safe Counterfactual Reinforcement Learning
Figure 3 for Safe Counterfactual Reinforcement Learning
Figure 4 for Safe Counterfactual Reinforcement Learning
Viaarxiv icon

Adaptive Experimental Design for Efficient Treatment Effect Estimation: Randomized Allocation via Contextual Bandit Algorithm

Add code
Feb 13, 2020
Figure 1 for Adaptive Experimental Design for Efficient Treatment Effect Estimation: Randomized Allocation via Contextual Bandit Algorithm
Figure 2 for Adaptive Experimental Design for Efficient Treatment Effect Estimation: Randomized Allocation via Contextual Bandit Algorithm
Figure 3 for Adaptive Experimental Design for Efficient Treatment Effect Estimation: Randomized Allocation via Contextual Bandit Algorithm
Figure 4 for Adaptive Experimental Design for Efficient Treatment Effect Estimation: Randomized Allocation via Contextual Bandit Algorithm
Viaarxiv icon

Efficient Counterfactual Learning from Bandit Feedback

Add code
Sep 10, 2018
Figure 1 for Efficient Counterfactual Learning from Bandit Feedback
Figure 2 for Efficient Counterfactual Learning from Bandit Feedback
Viaarxiv icon