Picture for Nathan Kallus

Nathan Kallus

Reward Maximization for Pure Exploration: Minimax Optimal Good Arm Identification for Nonparametric Multi-Armed Bandits

Add code
Oct 21, 2024
Figure 1 for Reward Maximization for Pure Exploration: Minimax Optimal Good Arm Identification for Nonparametric Multi-Armed Bandits
Figure 2 for Reward Maximization for Pure Exploration: Minimax Optimal Good Arm Identification for Nonparametric Multi-Armed Bandits
Figure 3 for Reward Maximization for Pure Exploration: Minimax Optimal Good Arm Identification for Nonparametric Multi-Armed Bandits
Figure 4 for Reward Maximization for Pure Exploration: Minimax Optimal Good Arm Identification for Nonparametric Multi-Armed Bandits
Viaarxiv icon

Adjusting Regression Models for Conditional Uncertainty Calibration

Add code
Sep 26, 2024
Viaarxiv icon

CSPI-MT: Calibrated Safe Policy Improvement with Multiple Testing for Threshold Policies

Add code
Aug 21, 2024
Viaarxiv icon

Estimating Heterogeneous Treatment Effects by Combining Weak Instruments and Observational Data

Add code
Jun 10, 2024
Viaarxiv icon

Contextual Linear Optimization with Bandit Feedback

Add code
May 26, 2024
Viaarxiv icon

Reindex-Then-Adapt: Improving Large Language Models for Conversational Recommendation

Add code
May 20, 2024
Viaarxiv icon

Demistifying Inference after Adaptive Experiments

Add code
May 02, 2024
Viaarxiv icon

Efficient and Sharp Off-Policy Evaluation in Robust Markov Decision Processes

Add code
Mar 29, 2024
Viaarxiv icon

Hessian-Free Laplace in Bayesian Deep Learning

Add code
Mar 15, 2024
Viaarxiv icon

Switching the Loss Reduces the Cost in Batch Reinforcement Learning

Add code
Mar 12, 2024
Figure 1 for Switching the Loss Reduces the Cost in Batch Reinforcement Learning
Figure 2 for Switching the Loss Reduces the Cost in Batch Reinforcement Learning
Figure 3 for Switching the Loss Reduces the Cost in Batch Reinforcement Learning
Figure 4 for Switching the Loss Reduces the Cost in Batch Reinforcement Learning
Viaarxiv icon