\floatsetup
[table]capposition=top
RuxinZheng
The University of Alabama
Tuscaloosa, AL 35487
&ShunqiaoSun
The University of Alabama
Tuscaloosa, AL 35487
HolgerCaesar
Delft University of Technology
Delft, Netherlands
&HongleiChen
Mathworks, Inc
Natick, MA 01760
&JianLi
University of Florida
Gainesville, FL 32611
Abstract
Millimeter-wave (mmWave) radars are indispensable for perception tasks of autonomous vehicles, thanks to their resilience in challenging weather conditions. Yet, their deployment is often limited by insufficient spatial resolution for precise semantic scene interpretation. Classical super-resolution techniques adapted from optical imaging inadequately address the distinct characteristics of radar signal data. In response, our study redefines radar imaging super-resolution as a one-dimensional (1D) signal super-resolution spectra estimation problem by harnessing the radar signal processing domain knowledge, introducing innovative data normalization and a domain-informed signal-to-noise ratio (SNR)-guided loss function. Our tailored deep learning network for automotive radar imaging exhibits remarkable scalability, parameter efficiency and fast inference speed, alongside enhanced performance in terms of radar imaging quality and resolution. Extensive testing confirms that our SR-SPECNet sets a new benchmark in producing high-resolution radar range-azimuth images, outperforming existing methods across varied antenna configurations and dataset sizes. Source code and new radar dataset will be made publicly available online.
1 Introduction
Radar technology, particularly in the form of millimeter wave radars, has become a cornerstone for advanced driver assistance systems (ADAS) and autonomous vehicles, surpassing the capabilities of traditional RGB cameras and LiDAR in challenging weather and low visibility conditions (23; 9; 34; 36; 24; 40). Its adoption is largely driven by the robust, cost-effective, and reliable sensing solutions it offers, operational under virtually all environmental scenarios. Frequency-modulated continuous-wave (FMCW) signals within the millimeter-wave band are primarily utilized in these radar systems, chosen for their cost-efficient operation and potential for high-resolution sensing. This technological choice is pivotal for a broad spectrum of autonomous driving functionalities, including free space detection, surrounding sensing, object detection and classification, and simultaneous localization and mapping (SLAM)(33; 8; 10).
Historically, automotive radar technology, dating back to the late 1990s and early 2000s, was developed with a focus on supporting ADAS functions like adaptive cruise control (ACC) (36). However, these radar systems primarily measure speed and range, offering limited azimuth angular resolution. To achieve Level 4 and Level 5 fully autonomous driving capabilities, a demand for high-resolution four-dimensional (4D) sensing has emerged(33). Such advanced sensing is essential not only for speed and range determination but also for accurately estimating targets’ azimuth and elevation with high resolution.
However, these radar systems primarily measure speed and range, offering limited azimuth angular resolution. The challenge of enhancing angular resolution has led to the extensive use of multiple-input multiple-output (MIMO) radar technology. MIMO radars synthesize a large virtual array aperture, significantly improving angular resolution with a manageable number of transmit and receive antennas (14; 15; 34; 3).
Further, signal processing techniques have been explored to further improve the angular resolution beyond what is achievable through digital beamforming technique that is implemented via fast Fourier transform (FFT). Super-resolution direction of arrival (DOA) estimation algorithms, such as compressive sensing (CS) (7; 6; 5) and the iterative adaptive approach (IAA) (39; 27), represent significant strides in this direction. Yet, their computational demand presents a formidable barrier to real-time implementation, especially in the dynamic context of automotive scenarios. Figure 1 illustrates how antenna aperture and super-resolution algorithms influence the quality of range-azimuth (RA) heatmaps. The high-resolution RA heatmaps contain rich information of the objects, including their shapes, facilitating object detection and classification through deep neural networks (40).
![Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (1) Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (1)](https://i0.wp.com/arxiv.org/html/2406.07399v1/x1.png)
The adoption of deep learning (DL) techniques for radar image enhancement has yielded significant advances within the realm of image super-resolution, as demonstrated in computer vision research (1; 11; 16). Applying these methods to enhance azimuth resolution in RA heatmaps presents a significant opportunity for substantial improvement. Nonetheless, few studies have focused on generating super-resolution RA heatmaps using raw radar signals by exploiting radar domain knowledge. Approaches that consider generating radar super-resolution RA heatmaps as straightforward image-to-image or volume-to-volume tasks often overlook the critical domain knowledge of radar signal processing. This oversight can lead to data-intensive solutions, rely on excessively large networks, or fail to deliver optimal performance and scalability which are key concerns in automotive applications where rapid inference and compact model size are essential for on-chip implementation.
Research in the domain of super-resolution RA heatmap generation for automotive radar remains limited, with the majority of studies relying on FFT-generated ground truths from larger antenna arrays. To our best knowledge, no existing methods leverage RA heatmaps produced through super-resolution algorithms as ground truths. Moreover, these methods typically focus on smaller antenna arrays and do not explore the potential of varied training data sizes. This paper aims to close these gaps by introducing the Super-Resolution Angular Spectra Estimation Network (SR-SPECNet). Designed with radar signal processing expertise, SR-SPECNet advances super-resolution angular spectra generation by transforming RA heatmap enhancement into a manageable 1D azimuth super-resolution challenge. This transformation is supported by our novel data normalization approach and an signal-to-noise ratio (SNR)-guided loss function. SR-SPECNet is thoroughly evaluated across varied antenna apertures and training dataset sizes, a first in this research area, using a dedicated real-world dataset. Our experimental analysis confirms that SR-SPECNet achieves exceptional parameter efficiency, superior performance in imaging quality, and outstanding scalability. It consistently surpasses established benchmarks, showcasing its capability to adapt to various antenna configurations and dataset sizes.
The key contributions of our work include:
- •
We introduced SR-SPECNet, a network designed for efficiency and effectiveness, capable of using single-snapshot measurement to robustly produce high-resolution automotive RA imaging typically obtained through IAA, but without IAA’s computational expense.
- •
We adopted radar signal processing domain knowledge to guide the neural network design by translating the RA imaging as a 1D spectra estimation problem, introducing a novel real radar data normalization method and a SNR-guided loss function.
- •
SR-SPECNet is the first network proven to robustly create high-resolution RA imaging with fast inference time from real automotive radar data featuring dynamic objects, demonstrating SR-SPECNet’s scalability, efficiency, and robust performance.
2 Related Work
The quest for enhanced radar imaging has predominantly focused on improving the azimuth resolution with limited number of antenna elements, given that range resolution can be augmented by increasing the bandwidth. In the automotive radar domain, digital beamforming (DBF) has emerged as the predominant DOA estimation algorithm, favored for its computational efficiency and robustness. This technique, typically implemented via FFT, however, faces limitations in angular resolution due to the Rayleigh criterion and is characterized by relatively high sidelobes (34; 26). Automotive radars, operating within highly dynamic environments, often have access to only a limited number of snapshots, sometimes as few as a single snapshot.This scenario renders super-resolution methods, such as Capon beamforming, MUSIC (30), and ESPRIT (29), which require multiple snapshots for covariance matrix estimation, less viable.
Compressive sensing (CS) techniques, which leverage the sparsity of target distributions in the angular domain, are highly effective in super-resolution, especially in snapshot-constrained settings (7; 5). Despite their potential, CS methods demand a dictionary matrix with low mutual coherence, which can be limiting. Alternatively, the Iterative Adaptive Approach (IAA) offers robust Direction of Arrival (DOA) estimation with limited snapshots, utilizing a nonparametric, iterative process (39; 27). IAA is advantageous for high-resolution radar imaging, surpassing subspace methods like MUSIC and ESPRIT and CS-based methods that falter under snapshot constraints or produce sparse results. However, IAA is computationally intensive, requiring large-scale matrix inversions at each step.Although fast and super-fast IAA variants (12; 38; 13) aim to reduce these demands by replacing matrix inversions with factorization, their benefits are marginal for small arrays and more pronounced for larger arrays, though computational challenges persist.
Recently, deep learning techniques have been applied to address the intricacies of azimuth super-resolution in RA maps.An adversarial network was tailored for super-resolution in micro-Doppler imagery (1), showcasing the potential of generative adversarial networks (GANs) in radar image enhancement. A U-Net architecture was employed for the super-resolution of weather radar maps (11), demonstrating the adaptability of deep convolutional networks to various radar data modalities. Notably, (16) ventured into extrapolating received antenna signals through a compact network, followed by the application of a 3D U-Net on the range-Doppler-azimuth data cube, facilitating the generation of super-resolution RA heatmaps.However, the existing body of work primarily leverages 2D or 3D network architectures, predicated on the assumption that the problem space necessitates multi-dimensional data processing to achieve enhanced resolution. This perspective, while valid, overlooks the potential efficiencies and novel insights that can be garnered from reinterpreting the challenge through a one-dimensional lens. To our best knowledge, no prior work has endeavored to address radar azimuth super-resolution within RA heatmaps using a 1D approach.In this paper, we close this gap by designing an efficient and effective deep neural network to achieving super-resolution radar imaging, through leveraging radar signal processing domain knowledge.
3 Radar Datasets
Dataset # of Frames Data Type Resolution Radar/Technology nuScenes (4) Sparse PC Low Continental ARS408 Oxford Radar (2) RA High Navtech Spinning Radar RADIATE (32) RA High Navtech Spinning Radar CRUW (37) RA Low TI AWR1843 Zendar (18) ADC,RD,PC High SAR CARRADA (20) RA,RD,RAD Low TI AWR1843 RadarScenes (31) Dense PC High 77GHz Middle-Range Radar RADIal(25) ADC,RD,PC High Valeo Middle Range DDM View-of-Delft(22) PC+Doppler High ZF FRGen21 Radar K-Radar (21) 4D Tensor High KAIST-Radar Radatron (17) 4D Tensor High TI Cascade Imaging Radar Ours ADC High TI Cascade Imaging Radar
Radar datasets for autonomous driving such as nuScenes (4), Oxford Radar RobotCar (2), RADIATE (32), and others, are summarized in Table 1.Technologies like spinning radar, utilized in the RADIATE and Oxford Radar RobotCar datasets, provide high-resolution 360-degree field-of-view (FOV) imagery, albeit at limited frame rates, which can introduce motion blur challenges. Dataset such as CARRADA employs single-chip Texas Instruments (TI) radar systems, offering modest angular resolutions exceeding .The Zendar dataset, leveraging synthetic aperture radar (SAR) technology, excels in imaging static targets by integrating measurements from different vehicle positions.The View-of-Delft dataset takes advantage of the ZF FRGen21 radar’s long-range and high-resolution imaging capabilities, offering point cloud data with object annotations confined to a 50 meter range.
3.1 Our Dataset
Our approach demands detailed radar configuration parameters and intensive raw analog-to-digital converter (ADC) data processing to integrate super-resolution algorithms effectively and assess network performance across various antenna apertures. Hence, we created our own dataset by driving a Lexus RX450h SUV equipped with multi-modal sensors, including a TI imaging radar, Teledyne FLIR Blackfly S stereo cameras, and a Velodyne Ultra Puck VLP-32C LiDAR sensor, along urban streets, highways and campus roads. The centerpiece of our dataset is the TI cascaded imaging radar system (35), configured for MIMO operations with an array of 12 transmit (TX) and 16 receive (RX) antennas.The operational 9 TX and 16 RX antennas were arranged to form a virtual uniform linear array (ULA) of 86 elements, with half-wavelength spacing, rendering an azimuth resolution of roughly 1.2 degrees via FFT.Our dataset showcases the exceptional high-resolution capabilities of our radar configuration, as illustrated in Figure 2.
![Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (2) Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (2)](https://i0.wp.com/arxiv.org/html/2406.07399v1/x2.png)
4 Method
We aim to transform raw ADC data into high-resolution RA maps. Unlike recent approaches that derive high-resolution ground truth from RA maps using an expanded antenna array (16), our method relies on RA maps generated with the same number of antennas but refined using IAA algorithms as our benchmark.
Figure 3 depicts our processing workflow. The input ADC data, , encapsulates three dimensions: for fast time samples, for slow time samples (or chirps), and for channels (or receivers). Through a 2D FFT to across both fast and slow time dimensions, we obtain range-Doppler-channel data, . Subsequently, beam vectors are extracted from each range-Doppler bin. These vectors are processed by SR-SPECNet to generate a super-resolution spectrum.This operation, performed on all beam vectors across all range-Doppler bins, yields the range-Doppler-azimuth data, . Notably, this procedure is highly parallelizable, treating the dataset as a 2D matrix with a batch size of . The final high-resolution RA maps, , are achieved by averaging over Doppler dimension.
![Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (3) Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (3)](https://i0.wp.com/arxiv.org/html/2406.07399v1/x3.png)
To provide a clearer understanding of the beam vector in the context of automotive radar signal processing, its signal model can be articulated as follows:
(1) |
where encapsulates the DOA of targets, signifies a complex white Gaussian noise vector, and represents the array manifold matrix for targets. The array response vector is given as . In this model, denotes the spacing between the -th element and the reference element, and represents the vector of source strengths.
4.1 SR-SPECNet for Azimuth Super-Resolution
CS and IAA stand out as widely recognized super-resolution DOA estimation algorithms tailored for single-snapshot radar data. In contrast to CS, which yields spectra consisting solely of discrete points, the IAA generates continuous spectra that include estimated reflection coefficients. This particular feature renders IAA more apt for the generation of RA maps.
As a super-resolution DOA estimation algorithm working under single-snapshot, IAA generates continuous spectra that include estimated reflection coefficients, rendering IAA more apt for the generation of RA maps.
4.1.1 Iterative Adaptive Approach (IAA)
IAA is a data-dependent, nonparametric algorithm (39). By discretizing the DOA space into an point grid, the array manifold is defined as with being the array steering vector. The fictitious covariance matrix of is represented as , where is a diagonal matrix with the -th diagonal element being , and is the source reflection coefficient corresponding to direction . IAA iteratively estimates the reflection coefficient and updates the fictitious covariance matrix by minimizing a weighted least-square (WLS) cost function , where and .
4.1.2 SR-SPECNet
SR-SPECNet’s primary objective is to transform input beam vectors into super-resolution IAA spectra. The IAA, functioning as an advanced beamforming algorithm, iteratively refines a reconstructed covariance matrix to estimate the spectrum as , where are the beamforming weights. At each beamforming angle , the array’s response, , parallels the output process of a multi-layer perceptron (MLP), underscoring the suitability of using an MLP for this application (19).
Designed as a four-layer MLP, SR-SPECNet mirrors the mathematical operations in the IAA algorithm. It processes the input signal, by separating its real and imaginary parts and concatenating them into a real-valued input. This approach ensures the preservation of crucial phase information, as complex value multiplication inherently involves both the real and imaginary parts.
4.2 Data Preprocessing
Proper data normalization is crucial for training neural networks, especially for regression tasks.Different from simulated signals with controlled factors like SNR, target reflection, and target number, real-world signals add unpredictability in SNR and reflections, challenging normalization. SNR varies significantly within a radar frame and from frame to frame. Maintaining a comparable intensity among beam vectors is crucial for constructing accurate RA heatmaps, where factors like SNR, target reflection intensity, and number of targets are precisely controlled. Real-world signals, however, introduce complexities not present in simulated environments: SNR and target reflection intensities vary unpredictably, complicating the task of normalizing inputs and labels. Within a single radar frame, the range of SNR values across beam vectors can be vast, with some vectors lacking targets altogether or presenting very low SNR. Adding to the complexity is the necessity to maintain the relative intensity among beam vectors within each radar frame, a critical aspect for accurately constructing the final RA heatmap.
To overcome the normalization challenges posed by the variability in real-world signals,we introduce a frequency domain normalization method designed to produce consistent and interpretable inputs for neural network training. This approach entails determining a normalization factor, , for each beam vector, calculated as the maximum absolute value of the frequency spectra, obtained by multiplying to the beam vector, equivalent to an FFT operation, and then divided by the total number of elements, :
(2) |
Subsequently, the raw signal is normalized using to yield , ensuring that the signal levels are stable across varying SNR conditions. Similarly, the label, represented by the IAA spectra , is normalized to . This normalization strategy effectively scales the signal and the IAA spectra so that their values fall within a comparable range, thereby facilitating more effective network training.Further, maintains a relative intensity among all beam vectors within each radar frame, ensuring that the spatial relationships that are critical for accurate RA heatmap synthesis are preserved.Moreover, preserves the relative intensity across all beam vectors in a radar frame, maintaining the spatial relationships essential for accurate synthesis of RA heatmaps. It’s important to note that this normalization process is exclusively needed during training and is not required for generating super-resolution RA heatmaps with test data, which stands as a significant advantage.This is attributed to the linear relationship between the beam vector and its corresponding spectra, allowing for direct processing without the need for normalization in the testing phase.
4.3 SNR-Guided Loss Function
The normalization factor , which represents the maximum value in the signal’s frequency domain, is directly proportional to the signal’s SNR. A higher suggests a higher SNR, positively influencing the quality of the final RA heatmap. We introduce an SNR-guided loss function similar to a weighted mean squared error (MSE), designed to prioritize higher SNR signals during training. The loss function is defined as:
where is the number of angle grid points of the spectra, and are the actual and predicted values at . This approach ensures that our model is finely tuned to emphasize higher quality signals.
5 Experiment
We train and evaluate our SR-SPECNet model using our own dataset, which comprises 17,000 frames of raw ADC radar data. To promote data diversity and minimize the redundancy of consecutive frames, we strategically selected every tenth frame from the dataset, yielding 1,700 frames, with the initial 1,400 frames dedicated to training the model, and the subsequent 300 frames reserved for testing. We intentionally structured the training frames into three subsets to simulate real-world data collection scenarios of limited time periods: a ‘small’ dataset with the initial 200 frames (akin to a 200-second data collection period), a ‘medium’ dataset comprising the first 700 frames, and a ‘large’ dataset that includes all 1,400 frames. This segmentation aims to test our model’s performance and adaptability under varying lengths of data availability.
5.1 Benchmarks
To evaluate SR-SPECNet’s effectiveness, we compare it with models designed to enhance spatial resolution. This comparison includes a 2D U-Net (28), which transforms low-resolution RA heatmaps into high-resolution equivalents, and the RAD-UNet (16), referred to as a 3D U-Net, that upgrades low-resolution range-azimuth-Doppler (RAD) data to high-resolution RAD imagery. We exclude pixel-based super-resolution networks like SRGAN, which increase resolution by adding pixels. These models do not meet the specific requirements of radar imaging, where resolution is not directly related to pixel count (16).
5.2 Evaluation Metrics
The RA map is a grayscale image, normalized between 0 to 1, for both generated and ground truth images. To comprehensively evaluate the quality of high-resolution RA heatmaps, we use established image evaluation metrics. PSNR, measured in dB, and SSIM, ranging from 0 to 1, assess image quality where higher values indicate better quality. NMSE also ranges from 0 to 1 and quantifies prediction accuracy by comparing the mean squared error to the variance of actual values, with lower values indicating more accurate predictions. Together, PSNR, SSIM, and NMSE provide a robust framework for assessing image fidelity, error magnitude, and compositional changes affecting perceived quality.
5.3 Implementation Details
Our radar configuration is characterized by fast-time samples , slow-time samples , and post-MIMO processing, resulting in beam vectors, each with elements.We truncated range of the radar data cube, , to keep the first elements along the range axis, resulting in a truncated dataset . This truncation strategy is informed by the observation that significant target information is concentrated within the first meters of the collected data.
We embark on an exploratory analysis of the effect of antenna aperture size on network performance. To this end, we selected a 10-element antenna array to represent a smaller aperture and a 40-element array for a larger aperture. The choice of a 40-element array as the larger aperture is strategically made, considering that it provides sufficient resolution, achieving approximately angular resolution using the IAA algorithm.
We set the angular grid size to for frequency domain uniformity. The labels for our 10-element and 40-element antenna arrays stem from their IAA spectra, ensuring our network’s performance evaluation remains consistent across varying apertures. SR-SPECNet comprises four fully connected layers, with the first three followed by ReLU activation functions and output sizes of 2048, 1024, and 512, respectively. The final layer’s output size matches . As input signals are complex, we concatenate their real and imaginary parts into a real-valued vector, serving as the input to SR-SPECNet.
We implemented SR-SPECNet and benchmark models using PyTorch, standardizing training with the Adam optimizer at a learning rate of 0.0001 for 500 epochs. Training was accelerated on four Nvidia RTX A6000 GPUs for efficiency.
5.4 High-Resolution RA Heatmap
We study the performance of deep neural networks in generating high-resolution RA heatmaps. In pursuit of this, we evaluated SR-SPECNet and SR-SPECNet+, which were trained with MSE loss and our SNR-guided loss, respectively, against established benchmark models. As delineated in Tables 2 and 3, our models were tested using 10 and 40 antenna elements across small, medium, and large dataset sizes.
For the 10-element antenna configuration, Table 2 highlights the strength of our methodology. SR-SPECNet+ emerges as the leading model, eclipsing both the 2D and 3D U-Net benchmarks as well as SR-SPECNet across all dataset sizes. Its edge in performance can be attributed to the incorporation of domain knowledge through the use of an SNR-guided loss function, which fine-tunes the training process to emphasize data with higher signal integrity. This advanced loss function allows SR-SPECNet+ to achieve the lowest NMSE and the highest SSIM and PSNR scores, demonstrating its effectiveness even in small dataset scenarios. SR-SPECNet itself outpaces conventional U-Net models, underscoring the value of generating RA heatmap through a 1D super-resolution lens. The comparative results make it evident that the domain-specific enhancements in SR-SPECNet+ significantly boost its ability to generate high-quality heatmaps, particularly when the number of antenna elements is limited.
Models small medium large NMSE SSIM PSNR NMSE SSIM PSNR NMSE SSIM PSNR 2D U-Net 0.321 0.782 27.780 0.233 0.820 29.182 0.104 0.877 32.736 3D U-Net 0.763 0.841 26.655 0.205 0.894 31.318 0.132 0.917 33.958 SR-SPECNet 0.168 0.909 30.683 0.092 0.946 33.446 0.077 0.955 34.326 SR-SPECNet+ 0.080 0.950 34.006 0.063 0.962 35.350 0.056 0.965 35.884
Shifting the focus to the 40-element antenna configuration, as presented in Table 3, SR-SPECNet+ maintains its exceptional standard. It substantially betters its NMSE and improves SSIM and PSNR values, asserting its robustness across all data volumes. SR-SPECNet retains a strong performance profile, lending weight to the concept that a 1D super-resolution approach effectively facilitates the production of high-quality heatmaps without extensive training data. This evidence underscores our model’s proficiency in handling data from varying antenna aperture sizes and confirms the strategic value of the 1D super-resolution methodology, which maximizes training data efficiency and is conducive to high-resolution radar imaging applications.
Models small medium large NMSE SSIM PSNR NMSE SSIM PSNR NMSE SSIM PSNR 2D U-Net 0.233 0.903 34.676 0.173 0.934 36.276 0.124 0.941 36.887 3D U-Net 1.255 0.813 31.649 1.320 0.816 33.106 0.292 0.946 37.378 SR-SPECNet 0.268 0.929 34.135 0.170 0.942 36.084 0.115 0.960 38.100 SR-SPECNet+ 0.144 0.951 36.924 0.122 0.959 37.916 0.093 0.965 39.066
5.5 Complexity and Scalability
The 2D and 3D U-Nets, designed for processing low-resolution RA and RAD heatmaps, have fixed numbers of trainable parameters for different size antenna arrays: approximately 31.0M for the 2D U-Net and 51.8M for the 3D U-Net. Their inference times are 6.14 ms and 6.98 ms, respectively, indicating the increased computational demands of higher-dimensional data processing. In contrast, our SR-SPECNet achieves notable efficiency, operating with just 1M parameters and a swift inference time of 3.12 ms. This reduction in time and model size enhances the speed and resource efficiency of our method, ideal for real-time automotive radar applications on embedded CPUs. Meanwhile, the IAA’s inference times are 12.3 ms for 10-element vectors and 30.73 ms for 40-element vectors, underscoring its computational challenges.
![Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (4) Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (4)](https://i0.wp.com/arxiv.org/html/2406.07399v1/x4.png)
Our SR-SPECNet model is designed for adaptability, accepting 1D beam vectors as input, which provides superior scalability across varying and values. As outlined in Section 5.3, our training dataset is configured with and . However, as demonstrated in Figure4, SR-SPECNet effortlessly handles different sizes of these parameters. We generate low-resolution and ground truth heatmaps using FFT and the IAA algorithm, respectively. This remarkable scalability highlights the advantages of our approach, confirming its suitability for diverse and dynamic radar imaging scenarios.
5.6 Generalizability
Our training and test datasets feature a rich variety of signals, with each radar frame containing thousands of signals to ensure diversity. To further test the model’s generalizability, we evaluated our pre-trained SR-SPECNet on 10,500 radar frames from Radartron (17), which present different and . Unlike the 2D and 3D UNets, which could not be applied directly due to these variations, our 1D approach was seamlessly implemented, showcasing its superior scalability. Performance metrics include a NMSE of 0.025, SSIM of 0.983, and PSNR of 37.858 for the 10-element configuration, and NMSE of 0.0743, SSIM of 0.970, and PSNR of 42.941 for the 40-element configuration.
![Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (5) Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (5)](https://i0.wp.com/arxiv.org/html/2406.07399v1/x5.png)
5.7 Visualization of RA maps
To evaluate the quality of high-resolution RA heatmap by SR-SPECNet and SR-SPECNet+, we visually compare the outputs against those from baseline models, i.e., 2D U-Net and 3D U-Net, as depicted in Fig. 5. Each model is trained on the ‘large’ dataset to obtain the best performance. Ground truth heatmaps generated with the IAA, have much better resolution than the LR images.Further, IAA suppresses sidelobes, yielding much clearer heatmaps. This underscores the benefits of using IAA-derived as ground truth for learning, rather than using FFT-generated heatmaps with a larger antenna aperture, which may not always be practically available due to hardware cost and the complexities involved in MIMO technology implementations.Fig. 5 demonstrates thatthe RA heatmaps that are respectively produced by SR-SPECNet and SR-SPECNet+ with a SNR-guided loss function, exhibit a noticeable improvement over the heatmaps generated by 2D and 3D U-Net, aligning with the quantitative metrics presented in Tables 2 and 3. These results collectively affirm the effectiveness of our proposed 1D methodologies over the 2D and 3D methods.
6 Conclusions
In this study, we have advanced automotive radar imaging by introducing SR-SPECNet, a novel 1D network leveraging IAA-generated RA heatmaps as ground truth and integrating a unique SNR-guided loss function for super-resolution RA heatmap generation. Our approach, emphasizing a 1D signal processing perspective, has demonstrated superior performance on real radar measurements in terms of automotive radar imaging quality, scalability, and efficiency across varying antenna configurations and dataset sizes. These contributions enhance the fidelity of radar imaging in autonomous vehicles. They also opens avenues for future research, especially with our commitment to sharing our own radar dataset and source code resources with the research community. This work underscores the potential of deep learning-enhanced radar processing in improving navigational safety and robust perception in autonomous vehicles.
7 Limitation & Future works
Our proposed method opens up possibilities for super-resolution imaging across four-dimensional (4D) radar parameters, range, Doppler, azimuth, and elevation. Extending our method to fully support 4D radar imaging represents a key direction for our future research. Currently, our network is optimized for ULA only; thus, adapting our model to accommodate arbitrary array geometries, including various sparse configurations and arrays affected by random antenna failures, is crucial. Additionally, we plan to enhance our network’s performance across diverse dynamic ranges and in scenarios characterized by extremely low SNR. These improvements are vital for advancing the robustness and applicability of our technology in real-world environments.
References
- Armanious etal. [2019]K.Armanious, S.Abdulatif, F.Aziz, U.Schneider, and B.Yang.An adversarial super-resolution remedy for radar design trade-offs.In 2019 27th European Signal Processing Conference (EUSIPCO), pages 1–5, 2019.doi: 10.23919/EUSIPCO.2019.8902510.
- Barnes etal. [2020]D.Barnes, M.Gadd, P.Murcutt, P.Newman, and I.Posner.The Oxford radar robotcar dataset: A radar extension to the oxford robotcar dataset.In Proc. IEEE International Conference on Robotics and Automation (ICRA), Paris, France, May 31-Oct. 31, 2020.
- Bergin and Guerci [2018]J.Bergin and J.R. Guerci.MIMO Radar: Theory and Application.Boston, MA, Artech House, 2018.
- Caesar etal. [2020]H.Caesar, V.Bankiti, A.H. Lang, S.Vora, V.E. Liong, Q.Xu, A.Krishnan, Y.Pan, G.Baldan, and O.Beijbom.nuScenes: A multimodal dataset for autonomous driving.IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 16-18, 2020.
- Candes and Fernandez-Granda [2014]E.Candes and C.Fernandez-Granda.Towards a mathematical theory of super-resolution.Communications on Pure and Applied Mathematics, 67(6):906–956, 2014.
- Candès and Romberg [2007]E.Candès and J.Romberg.Sparsity and incoherence in compressive sampling.Inverse problems, 23(3):969, 2007.
- Donoho [2006]D.L. Donoho.Compressed sensing.IEEE Transactions on Information Theory, 52(4):1289–1306, 2006.
- Duggal etal. [2020]G.Duggal, S.Vishwakarma, K.V. Mishra, and S.S. Ram.Doppler-resilient 802.11ad-based ultrashort range automotive joint radar-communications system.IEEE Transactions on Aerospace and Electronic Systems, 56(5):4035–4048, 2020.
- Engels etal. [2017]F.Engels, P.Heidenreich, A.M. Zoubir, F.K. Jondral, and M.Wintermantel.Advances in automotive radar: A framework on computationally efficient high-resolution frequency estimation.IEEE Signal Processing Magazine, 34(2):36–46, 2017.
- Engels etal. [2021]F.Engels, P.Heidenreich, M.Wintermantel, L.Stäcker, M.AlKadi, and A.M. Zoubir.Automotive radar signal processing: Research directions and practical challenges.IEEE Journal of Selected Topics in Signal Processing, 15(4):865–878, 2021.
- Geiss and Hardin [2020]A.Geiss and J.C. Hardin.Radar super resolution using a deep convolutional neural network.Journal of Atmospheric and Oceanic Technology, 37(12):2197–2207, 2020.
- Glentis and Jakobsson [2011]G.-O. Glentis and A.Jakobsson.Efficient implementation of iterative adaptive approach spectral estimation techniques.IEEE Transactions on Signal Processing, 59(9):4154–4167, 2011.doi: 10.1109/TSP.2011.2145376.
- Glentis and Jakobsson [2012]G.O. Glentis and A.Jakobsson.Superfast approximative implementation of the IAA spectral estimate.IEEE Transactions on Signal Processing, 60(1):472–478, 2012.doi: 10.1109/TSP.2011.2170979.
- Li and Stoica [2007]J.Li and P.Stoica.MIMO radar with colocated antennas.IEEE Signal Processing Magazine, 24(5):106–114, 2007.doi: 10.1109/MSP.2007.904812.
- Li and Stoica [2009]J.Li and P.Stoica.MIMO Radar Signal Processing.Hoboken, NJ, Wiley, 2009.
- Li etal. [2023]Y.-J. Li, S.Hunt, J.Park, M.O’Toole, and K.Kitani.Azimuth super-resolution for FMCW radar in autonomous driving.In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 17504–17513, 2023.
- Madani etal. [2022]S.Madani, J.Guan, W.Ahmed, S.Gupta, and H.Hassanieh.Radatron: Accurate detection using multi-resolution cascaded MIMO radar.In European Conference on Computer Vision (ECCV), pages 160–178, 2022.
- Mostajabi etal. [2020]M.Mostajabi, C.M. Wang, D.Ranjan, and G.Hsyu.High resolution radar dataset for semi-supervised learning of dynamic objects.In Proc. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Seattle, WA, June 14-19, 2020.
- Naumovski and Carrasco [1995]M.Naumovski and R.Carrasco.Neural network beamformer for narrow-band HF transmission.In IEE Colloquium on HF Antennas and Propagation, pages 5/1–5/8, 1995.doi: 10.1049/ic:19951273.
- Ouaknine etal. [2021]A.Ouaknine, A.Newson, J.Rebut, F.Tupin, and P.Perez.CARRADA dataset: Camera and automotive radar with range-angle-Doppler annotations.In Proc. 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, Jan. 10-15, 2021.
- Paek etal. [2022]D.-H. Paek, S.-H. Kong, and K.T. Wijaya.K-radar: 4D radar object detection for autonomous driving in various weather conditions.In Thirty-sixth Conference on Neural Information Processing Systems Datasets and Benchmarks Track, 2022.URL https://openreview.net/forum?id=W_bsDmzwaZ7.
- Palffy etal. [2022]A.Palffy, E.Pool, S.Baratam, J.Kooij, and D.Gavrila.Multi-class road user detection with 3+1D radar in the view-of-Delft dataset.IEEE Robotics and Automation Letters, 7(2):4961–4968, 2022.doi: 10.1109/LRA.2022.3147324.
- Patole etal. [2017]S.Patole, M.Torlak, D.Wang, and M.Ali.Automotive radars: A review of signal processing techniques.IEEE Signal Processing Magazine, 34(2):22–35, 2017.
- Peng etal. [2022]Z.Peng, C.Li, and F.Uysal.Modern Radar for Automotive Applications.London, UK: IET, 2022.
- Rebut etal. [2022]J.Rebut, A.Ouaknine, W.Malik, and P.Pérez.Raw high-definition radar for multi-task learning.In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 17021–17030, 2022.
- Richards [2022]M.A. Richards.Fundamentals of Radar Signal Processing, 3rd Ed.New York, McGraw-Hill, 2022.
- Roberts etal. [2010]W.Roberts, P.Stoica, J.Li, T.Yardibi, and F.Sadjadi.Iterative adaptive approaches to MIMO radar imaging.4(1):5–20, 2010.
- Ronneberger etal. [2015]O.Ronneberger, P.Fischer, and T.Brox.U-net: Convolutional networks for biomedical image segmentation.In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18, pages 234–241. Springer, 2015.
- Roy and Kailath [1989]R.Roy and T.Kailath.ESPRIT-estimation of signal parameters via rotational invariance techniques.IEEE Transactions on Acoustics, Speech, and Signal Processing, 37(7):984–995, 1989.doi: 10.1109/29.32276.
- Schmidt [1982]R.O. Schmidt.A signal subspace approach to multiple emitter location and spectral estimation.Stanford University, 1982.
- Schumann etal. [2021]O.Schumann, M.Hahn, N.Scheiner, F.Weishaupt, J.F. Tilly, J.Dickmann, and C.Wöhler.RadarScenes: A real-world radar point cloud data set for automotive applications.arXiv preprint arXiv:2104.02493, 2021.
- Sheeny etal. [2020]M.Sheeny, E.D. Pellegrin, M.Saptarshi, A.Ahrabian, S.Wang, and A.Wallace.RADIATE: A radar dataset for automotive perception.arXiv preprint arXiv:2010.09076, 2020.
- Sun and Zhang [2021]S.Sun and Y.D. Zhang.4D automotive radar sensing for autonomous vehicles: A sparsity-oriented approach.IEEE Journal of Selected Topics in Signal Processing, 15(4):879–891, 2021.
- Sun etal. [2020]S.Sun, A.P. Petropulu, and H.V. Poor.MIMO radar for advanced driver-assistance systems and autonomous driving: Advantages and challenges.IEEE Signal Processing Magazine, 37(4):98–117, 2020.
- Texas Instruments [2020]Texas Instruments.Design Guide: TIDEP-01012 Imaging Radar Using Cascaded MmWave Sensor Reference Design, 2020.URL https://www.ti.com/lit/ug/tiduen5a/tiduen5a.pdf.Rev. A.
- Waldschmidt etal. [2021]C.Waldschmidt, J.Hasch, and W.Menzel.Automotive radar — From first efforts to future systems.IEEE Journal of Microwaves, 1(1):135–148, 2021.
- Wang etal. [2021]Y.Wang, Z.Jiang, X.Gao, J.-N. Hwang, G.Xing, and H.Liu.RODNet: Radar object detection using cross-modal supervision.In Proc. IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, Jan. 5-9, 2021.
- Xue etal. [2011]M.Xue, L.Xu, and J.Li.IAA spectral estimation: Fast implementation using the Gohberg-sem*ncul factorization.IEEE Transactions on Signal Processing, 59(7):3251–3261, 2011.doi: 10.1109/TSP.2011.2131136.
- Yardibi etal. [2010]T.Yardibi, J.Li, P.Stoica, M.Xue, and A.Baggeroer.Source localization and sensing: A nonparametric iterative adaptive approach based on weighted least squares.46(1):425–443, 2010.
- Zheng etal. [2023]R.Zheng, S.Sun, H.Liu, and T.Wu.Deep-neural-network-enabled vehicle detection using high-resolution automotive radar imaging.IEEE Transactions on Aerospace and Electronic Systems, 59(5):4815–4830, 2023.