Picture for Alain Andres

Alain Andres

On the Inherent Robustness of One-Stage Object Detection against Out-of-Distribution Data

Add code
Nov 07, 2024
Viaarxiv icon

On the Black-box Explainability of Object Detection Models for Safe and Trustworthy Industrial Applications

Add code
Oct 28, 2024
Viaarxiv icon

Words as Beacons: Guiding RL Agents with High-Level Language Prompts

Add code
Oct 11, 2024
Figure 1 for Words as Beacons: Guiding RL Agents with High-Level Language Prompts
Figure 2 for Words as Beacons: Guiding RL Agents with High-Level Language Prompts
Figure 3 for Words as Beacons: Guiding RL Agents with High-Level Language Prompts
Figure 4 for Words as Beacons: Guiding RL Agents with High-Level Language Prompts
Viaarxiv icon

Fostering Intrinsic Motivation in Reinforcement Learning with Pretrained Foundation Models

Add code
Oct 09, 2024
Viaarxiv icon

Surgical Task Automation Using Actor-Critic Frameworks and Self-Supervised Imitation Learning

Add code
Sep 04, 2024
Figure 1 for Surgical Task Automation Using Actor-Critic Frameworks and Self-Supervised Imitation Learning
Figure 2 for Surgical Task Automation Using Actor-Critic Frameworks and Self-Supervised Imitation Learning
Figure 3 for Surgical Task Automation Using Actor-Critic Frameworks and Self-Supervised Imitation Learning
Figure 4 for Surgical Task Automation Using Actor-Critic Frameworks and Self-Supervised Imitation Learning
Viaarxiv icon

Enhanced Generalization through Prioritization and Diversity in Self-Imitation Reinforcement Learning over Procedural Environments with Sparse Rewards

Add code
Nov 01, 2023
Viaarxiv icon

Using Offline Data to Speed-up Reinforcement Learning in Procedurally Generated Environments

Add code
Apr 18, 2023
Figure 1 for Using Offline Data to Speed-up Reinforcement Learning in Procedurally Generated Environments
Figure 2 for Using Offline Data to Speed-up Reinforcement Learning in Procedurally Generated Environments
Figure 3 for Using Offline Data to Speed-up Reinforcement Learning in Procedurally Generated Environments
Figure 4 for Using Offline Data to Speed-up Reinforcement Learning in Procedurally Generated Environments
Viaarxiv icon

Towards Improving Exploration in Self-Imitation Learning using Intrinsic Motivation

Add code
Nov 30, 2022
Viaarxiv icon

An Evaluation Study of Intrinsic Motivation Techniques applied to Reinforcement Learning over Hard Exploration Environments

Add code
May 23, 2022
Figure 1 for An Evaluation Study of Intrinsic Motivation Techniques applied to Reinforcement Learning over Hard Exploration Environments
Figure 2 for An Evaluation Study of Intrinsic Motivation Techniques applied to Reinforcement Learning over Hard Exploration Environments
Figure 3 for An Evaluation Study of Intrinsic Motivation Techniques applied to Reinforcement Learning over Hard Exploration Environments
Figure 4 for An Evaluation Study of Intrinsic Motivation Techniques applied to Reinforcement Learning over Hard Exploration Environments
Viaarxiv icon

Collaborative Training of Heterogeneous Reinforcement Learning Agents in Environments with Sparse Rewards: What and When to Share?

Add code
Feb 24, 2022
Figure 1 for Collaborative Training of Heterogeneous Reinforcement Learning Agents in Environments with Sparse Rewards: What and When to Share?
Figure 2 for Collaborative Training of Heterogeneous Reinforcement Learning Agents in Environments with Sparse Rewards: What and When to Share?
Figure 3 for Collaborative Training of Heterogeneous Reinforcement Learning Agents in Environments with Sparse Rewards: What and When to Share?
Figure 4 for Collaborative Training of Heterogeneous Reinforcement Learning Agents in Environments with Sparse Rewards: What and When to Share?
Viaarxiv icon