Picture for Aadirupa Saha

Aadirupa Saha

Hybrid Preference Optimization for Alignment: Provably Faster Convergence Rates by Combining Offline Preferences with Online Exploration

Add code
Dec 13, 2024
Viaarxiv icon

Strategic Linear Contextual Bandits

Add code
Jun 01, 2024
Viaarxiv icon

DP-Dueling: Learning from Preference Feedback without Compromising User Privacy

Add code
Mar 22, 2024
Viaarxiv icon

Stop Relying on No-Choice and Do not Repeat the Moves: Optimal, Efficient and Practical Algorithms for Assortment Optimization

Add code
Feb 29, 2024
Viaarxiv icon

Think Before You Duel: Understanding Complexities of Preference Learning under Constrained Resources

Add code
Dec 28, 2023
Viaarxiv icon

Faster Convergence with Multiway Preferences

Add code
Dec 19, 2023
Viaarxiv icon

Federated Online and Bandit Convex Optimization

Add code
Nov 29, 2023
Viaarxiv icon

Bandits Meet Mechanism Design to Combat Clickbait in Online Recommendation

Add code
Nov 27, 2023
Figure 1 for Bandits Meet Mechanism Design to Combat Clickbait in Online Recommendation
Viaarxiv icon

Dueling Optimization with a Monotone Adversary

Add code
Nov 18, 2023
Viaarxiv icon

On the Vulnerability of Fairness Constrained Learning to Malicious Noise

Add code
Jul 26, 2023
Viaarxiv icon