Picture for Aaron Courville

Aaron Courville

Universite de Montreal

Stable Gradients for Stable Learning at Scale in Deep Reinforcement Learning

Add code
Jun 18, 2025
Viaarxiv icon

Adaptive Accompaniment with ReaLchords

Add code
Jun 17, 2025
Viaarxiv icon

The Courage to Stop: Overcoming Sunk Cost Fallacy in Deep Reinforcement Learning

Add code
Jun 16, 2025
Viaarxiv icon

Bias Analysis in Unconditional Image Generative Models

Add code
Jun 10, 2025
Viaarxiv icon

Measure gradients, not activations! Enhancing neuronal activity in deep reinforcement learning

Add code
May 29, 2025
Viaarxiv icon

FLAM: Frame-Wise Language-Audio Modeling

Add code
May 08, 2025
Viaarxiv icon

Adaptive Computation Pruning for the Forgetting Transformer

Add code
Apr 09, 2025
Viaarxiv icon

Forgetting Transformer: Softmax Attention with a Forget Gate

Add code
Mar 03, 2025
Viaarxiv icon

Asynchronous RLHF: Faster and More Efficient Off-Policy RL for Language Models

Add code
Oct 23, 2024
Figure 1 for Asynchronous RLHF: Faster and More Efficient Off-Policy RL for Language Models
Figure 2 for Asynchronous RLHF: Faster and More Efficient Off-Policy RL for Language Models
Figure 3 for Asynchronous RLHF: Faster and More Efficient Off-Policy RL for Language Models
Figure 4 for Asynchronous RLHF: Faster and More Efficient Off-Policy RL for Language Models
Viaarxiv icon

Stick-breaking Attention

Add code
Oct 23, 2024
Figure 1 for Stick-breaking Attention
Figure 2 for Stick-breaking Attention
Figure 3 for Stick-breaking Attention
Figure 4 for Stick-breaking Attention
Viaarxiv icon