Picture for Behzad Khamidehi

Behzad Khamidehi

Learning from Mistakes: a Weakly-supervised Method for Mitigating the Distribution Shift in Autonomous Vehicle Planning

Add code
Jun 03, 2024
Viaarxiv icon

Augmenting Safety-Critical Driving Scenarios while Preserving Similarity to Expert Trajectories

Add code
Apr 20, 2024
Viaarxiv icon

Distributed Deep Reinforcement Learning for Intelligent Traffic Monitoring with a Team of Aerial Robots

Add code
Jul 10, 2021
Figure 1 for Distributed Deep Reinforcement Learning for Intelligent Traffic Monitoring with a Team of Aerial Robots
Figure 2 for Distributed Deep Reinforcement Learning for Intelligent Traffic Monitoring with a Team of Aerial Robots
Figure 3 for Distributed Deep Reinforcement Learning for Intelligent Traffic Monitoring with a Team of Aerial Robots
Figure 4 for Distributed Deep Reinforcement Learning for Intelligent Traffic Monitoring with a Team of Aerial Robots
Viaarxiv icon

Dynamic Resource Management for Providing QoS in Drone Delivery Systems

Add code
Mar 06, 2021
Figure 1 for Dynamic Resource Management for Providing QoS in Drone Delivery Systems
Figure 2 for Dynamic Resource Management for Providing QoS in Drone Delivery Systems
Figure 3 for Dynamic Resource Management for Providing QoS in Drone Delivery Systems
Figure 4 for Dynamic Resource Management for Providing QoS in Drone Delivery Systems
Viaarxiv icon

Federated Learning for Cellular-connected UAVs: Radio Mapping and Path Planning

Add code
Aug 23, 2020
Figure 1 for Federated Learning for Cellular-connected UAVs: Radio Mapping and Path Planning
Figure 2 for Federated Learning for Cellular-connected UAVs: Radio Mapping and Path Planning
Figure 3 for Federated Learning for Cellular-connected UAVs: Radio Mapping and Path Planning
Figure 4 for Federated Learning for Cellular-connected UAVs: Radio Mapping and Path Planning
Viaarxiv icon

A Double Q-Learning Approach for Navigation of Aerial Vehicles with Connectivity Constraint

Add code
Feb 24, 2020
Figure 1 for A Double Q-Learning Approach for Navigation of Aerial Vehicles with Connectivity Constraint
Figure 2 for A Double Q-Learning Approach for Navigation of Aerial Vehicles with Connectivity Constraint
Figure 3 for A Double Q-Learning Approach for Navigation of Aerial Vehicles with Connectivity Constraint
Figure 4 for A Double Q-Learning Approach for Navigation of Aerial Vehicles with Connectivity Constraint
Viaarxiv icon

Reinforcement Learning-Based Trajectory Design for the Aerial Base Stations

Add code
Jun 29, 2019
Figure 1 for Reinforcement Learning-Based Trajectory Design for the Aerial Base Stations
Figure 2 for Reinforcement Learning-Based Trajectory Design for the Aerial Base Stations
Figure 3 for Reinforcement Learning-Based Trajectory Design for the Aerial Base Stations
Viaarxiv icon