Picture for Hengli Wang

Hengli Wang

Carl-Lead: Lidar-based End-to-End Autonomous Driving with Contrastive Deep Reinforcement Learning

Add code
Sep 17, 2021
Figure 1 for Carl-Lead: Lidar-based End-to-End Autonomous Driving with Contrastive Deep Reinforcement Learning
Figure 2 for Carl-Lead: Lidar-based End-to-End Autonomous Driving with Contrastive Deep Reinforcement Learning
Figure 3 for Carl-Lead: Lidar-based End-to-End Autonomous Driving with Contrastive Deep Reinforcement Learning
Figure 4 for Carl-Lead: Lidar-based End-to-End Autonomous Driving with Contrastive Deep Reinforcement Learning
Viaarxiv icon

Graph Attention Layer Evolves Semantic Segmentation for Road Pothole Detection: A Benchmark and Algorithms

Add code
Sep 06, 2021
Figure 1 for Graph Attention Layer Evolves Semantic Segmentation for Road Pothole Detection: A Benchmark and Algorithms
Figure 2 for Graph Attention Layer Evolves Semantic Segmentation for Road Pothole Detection: A Benchmark and Algorithms
Figure 3 for Graph Attention Layer Evolves Semantic Segmentation for Road Pothole Detection: A Benchmark and Algorithms
Figure 4 for Graph Attention Layer Evolves Semantic Segmentation for Road Pothole Detection: A Benchmark and Algorithms
Viaarxiv icon

DQ-GAT: Towards Safe and Efficient Autonomous Driving with Deep Q-Learning and Graph Attention Networks

Add code
Aug 11, 2021
Figure 1 for DQ-GAT: Towards Safe and Efficient Autonomous Driving with Deep Q-Learning and Graph Attention Networks
Figure 2 for DQ-GAT: Towards Safe and Efficient Autonomous Driving with Deep Q-Learning and Graph Attention Networks
Figure 3 for DQ-GAT: Towards Safe and Efficient Autonomous Driving with Deep Q-Learning and Graph Attention Networks
Figure 4 for DQ-GAT: Towards Safe and Efficient Autonomous Driving with Deep Q-Learning and Graph Attention Networks
Viaarxiv icon

SNE-RoadSeg+: Rethinking Depth-Normal Translation and Deep Supervision for Freespace Detection

Add code
Jul 30, 2021
Figure 1 for SNE-RoadSeg+: Rethinking Depth-Normal Translation and Deep Supervision for Freespace Detection
Figure 2 for SNE-RoadSeg+: Rethinking Depth-Normal Translation and Deep Supervision for Freespace Detection
Figure 3 for SNE-RoadSeg+: Rethinking Depth-Normal Translation and Deep Supervision for Freespace Detection
Figure 4 for SNE-RoadSeg+: Rethinking Depth-Normal Translation and Deep Supervision for Freespace Detection
Viaarxiv icon

Vision-Based Autonomous Car Racing Using Deep Imitative Reinforcement Learning

Add code
Jul 18, 2021
Figure 1 for Vision-Based Autonomous Car Racing Using Deep Imitative Reinforcement Learning
Figure 2 for Vision-Based Autonomous Car Racing Using Deep Imitative Reinforcement Learning
Figure 3 for Vision-Based Autonomous Car Racing Using Deep Imitative Reinforcement Learning
Figure 4 for Vision-Based Autonomous Car Racing Using Deep Imitative Reinforcement Learning
Viaarxiv icon

SCV-Stereo: Learning Stereo Matching from a Sparse Cost Volume

Add code
Jul 17, 2021
Figure 1 for SCV-Stereo: Learning Stereo Matching from a Sparse Cost Volume
Figure 2 for SCV-Stereo: Learning Stereo Matching from a Sparse Cost Volume
Figure 3 for SCV-Stereo: Learning Stereo Matching from a Sparse Cost Volume
Figure 4 for SCV-Stereo: Learning Stereo Matching from a Sparse Cost Volume
Viaarxiv icon

Co-Teaching: An Ark to Unsupervised Stereo Matching

Add code
Jul 17, 2021
Figure 1 for Co-Teaching: An Ark to Unsupervised Stereo Matching
Figure 2 for Co-Teaching: An Ark to Unsupervised Stereo Matching
Figure 3 for Co-Teaching: An Ark to Unsupervised Stereo Matching
Figure 4 for Co-Teaching: An Ark to Unsupervised Stereo Matching
Viaarxiv icon

End-to-End Interactive Prediction and Planning with Optical Flow Distillation for Autonomous Driving

Add code
Apr 18, 2021
Figure 1 for End-to-End Interactive Prediction and Planning with Optical Flow Distillation for Autonomous Driving
Figure 2 for End-to-End Interactive Prediction and Planning with Optical Flow Distillation for Autonomous Driving
Figure 3 for End-to-End Interactive Prediction and Planning with Optical Flow Distillation for Autonomous Driving
Figure 4 for End-to-End Interactive Prediction and Planning with Optical Flow Distillation for Autonomous Driving
Viaarxiv icon

Learning Interpretable End-to-End Vision-Based Motion Planning for Autonomous Driving with Optical Flow Distillation

Add code
Apr 18, 2021
Figure 1 for Learning Interpretable End-to-End Vision-Based Motion Planning for Autonomous Driving with Optical Flow Distillation
Figure 2 for Learning Interpretable End-to-End Vision-Based Motion Planning for Autonomous Driving with Optical Flow Distillation
Figure 3 for Learning Interpretable End-to-End Vision-Based Motion Planning for Autonomous Driving with Optical Flow Distillation
Figure 4 for Learning Interpretable End-to-End Vision-Based Motion Planning for Autonomous Driving with Optical Flow Distillation
Viaarxiv icon

S2P2: Self-Supervised Goal-Directed Path Planning Using RGB-D Data for Robotic Wheelchairs

Add code
Mar 18, 2021
Figure 1 for S2P2: Self-Supervised Goal-Directed Path Planning Using RGB-D Data for Robotic Wheelchairs
Figure 2 for S2P2: Self-Supervised Goal-Directed Path Planning Using RGB-D Data for Robotic Wheelchairs
Figure 3 for S2P2: Self-Supervised Goal-Directed Path Planning Using RGB-D Data for Robotic Wheelchairs
Figure 4 for S2P2: Self-Supervised Goal-Directed Path Planning Using RGB-D Data for Robotic Wheelchairs
Viaarxiv icon