Abstract:"This work has been submitted to the IEEE for possible publication. Copyright may be transferred without notice, after which this version may no longer be accessible." Driver's interaction with a vehicle via automatic gesture recognition is expected to enhance driving safety by decreasing driver's distraction. Optical and infrared-based gesture recognition systems are limited by occlusions, poor lighting, and varying thermal conditions and, therefore, have limited performance in practical in-cabin applications. Radars are insensitive to lighting or thermal conditions and, therefore, are more suitable for in-cabin applications. However, the spatial resolution of conventional radars is insufficient for accurate gesture recognition. The main objective of this research is to derive an accurate gesture recognition approach using low-resolution radars with deep learning-based super-resolution processing. The main idea is to reconstruct high-resolution information from the radar's low-resolution measurements. The major challenge is the derivation of the real-time processing approach. The proposed approach combines conventional signal processing and deep learning methods. The radar echoes are arranged in 3D data cubes and processed using a super-resolution model to enhance range and Doppler resolution. The FFT is used to generate the range-Doppler maps, which enter the deep neural network for efficient gesture recognition. The preliminary results demonstrated the proposed approach's efficiency in achieving high gesture recognition performance using conventional low-resolution radars.
Abstract:Vision-based autonomous driving requires reliable and efficient object detection. This work proposes a DiffusionDet-based framework that exploits data fusion from the monocular camera and depth sensor to provide the RGB and depth (RGB-D) data. Within this framework, ground truth bounding boxes are randomly reshaped as part of the training phase, allowing the model to learn the reverse diffusion process of noise addition. The system methodically enhances a randomly generated set of boxes at the inference stage, guiding them toward accurate final detections. By integrating the textural and color features from RGB images with the spatial depth information from the LiDAR sensors, the proposed framework employs a feature fusion that substantially enhances object detection of automotive targets. The $2.3$ AP gain in detecting automotive targets is achieved through comprehensive experiments using the KITTI dataset. Specifically, the improved performance of the proposed approach in detecting small objects is demonstrated.
Abstract:Autonomous driving and advanced active safety features require accurate high-resolution sensing capabilities. Automotive radars are the key component of the vehicle sensing suit. However, when these radars operate in proximity to flat surfaces, such as roads and guardrails, they experience a multipath phenomenon that can degrade the accuracy of the direction-of-arrival (DOA) estimation. Presence of multipath leads to misspecification in the radar data model, resulting in estimation performance degradation, which cannot be reliably predicted by conventional performance bounds. In this paper, the misspecified Cram\'er-Rao bound (MCRB), which accounts for model misspecification, is derived for the problem of DOA estimation in the presence of multipath which is ignored by the estimator. Analytical relations between the MCRB and the Cram\'er-Rao bound are established, and the DOA estimation performance degradation due to multipath is investigated. The results show that the MCRB reliably predicts the asymptotic performance of the misspecified maximum-likelihood estimator and therefore, can serve as an efficient tool for automotive radar performance evaluation and system design.
Abstract:This work addresses the problem of direction-of-arrival (DOA) estimation in the presence of non-Gaussian, heavy-tailed, and spatially-colored interference. Conventionally, the interference is considered to be Gaussian-distributed and spatially white. However, in practice, this assumption is not guaranteed, which results in degraded DOA estimation performance. Maximum likelihood DOA estimation in the presence of non-Gaussian and spatially colored interference is computationally complex and not practical. Therefore, this work proposes a neural network (NN) based DOA estimation approach for spatial spectrum estimation in multi-source scenarios with a-priori unknown number of sources in the presence of non-Gaussian spatially-colored interference. The proposed approach utilizes a single NN instance for simultaneous source enumeration and DOA estimation. It is shown via simulations that the proposed approach significantly outperforms conventional and NN-based approaches in terms of probability of resolution, estimation accuracy, and source enumeration accuracy in conditions of low SIR, small sample support, and when the angular separation between the source DOAs and the spatially-colored interference is small.
Abstract:While camera and LiDAR processing have been revolutionized since the introduction of deep learning, radar processing still relies on classical tools. In this paper, we introduce a deep learning approach for radar processing, working directly with the radar complex data. To overcome the lack of radar labeled data, we rely in training only on the radar calibration data and introduce new radar augmentation techniques. We evaluate our method on the radar 4D detection task and demonstrate superior performance compared to the classical approaches while keeping real-time performance. Applying deep learning on radar data has several advantages such as eliminating the need for an expensive radar calibration process each time and enabling classification of the detected objects with almost zero-overhead.