Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (2024)

\floatsetup

[table]capposition=top

RuxinZheng
The University of Alabama
Tuscaloosa, AL 35487
&ShunqiaoSun
The University of Alabama
Tuscaloosa, AL 35487
HolgerCaesar
Delft University of Technology
Delft, Netherlands
&HongleiChen
Mathworks, Inc
Natick, MA 01760
&JianLi
University of Florida
Gainesville, FL 32611

Abstract

Millimeter-wave (mmWave) radars are indispensable for perception tasks of autonomous vehicles, thanks to their resilience in challenging weather conditions. Yet, their deployment is often limited by insufficient spatial resolution for precise semantic scene interpretation. Classical super-resolution techniques adapted from optical imaging inadequately address the distinct characteristics of radar signal data. In response, our study redefines radar imaging super-resolution as a one-dimensional (1D) signal super-resolution spectra estimation problem by harnessing the radar signal processing domain knowledge, introducing innovative data normalization and a domain-informed signal-to-noise ratio (SNR)-guided loss function. Our tailored deep learning network for automotive radar imaging exhibits remarkable scalability, parameter efficiency and fast inference speed, alongside enhanced performance in terms of radar imaging quality and resolution. Extensive testing confirms that our SR-SPECNet sets a new benchmark in producing high-resolution radar range-azimuth images, outperforming existing methods across varied antenna configurations and dataset sizes. Source code and new radar dataset will be made publicly available online.

1 Introduction

Radar technology, particularly in the form of millimeter wave radars, has become a cornerstone for advanced driver assistance systems (ADAS) and autonomous vehicles, surpassing the capabilities of traditional RGB cameras and LiDAR in challenging weather and low visibility conditions (23; 9; 34; 36; 24; 40). Its adoption is largely driven by the robust, cost-effective, and reliable sensing solutions it offers, operational under virtually all environmental scenarios. Frequency-modulated continuous-wave (FMCW) signals within the millimeter-wave band are primarily utilized in these radar systems, chosen for their cost-efficient operation and potential for high-resolution sensing. This technological choice is pivotal for a broad spectrum of autonomous driving functionalities, including free space detection, 360superscript360360^{\circ}360 start_POSTSUPERSCRIPT ∘ end_POSTSUPERSCRIPT surrounding sensing, object detection and classification, and simultaneous localization and mapping (SLAM)(33; 8; 10).

Historically, automotive radar technology, dating back to the late 1990s and early 2000s, was developed with a focus on supporting ADAS functions like adaptive cruise control (ACC) (36). However, these radar systems primarily measure speed and range, offering limited azimuth angular resolution. To achieve Level 4 and Level 5 fully autonomous driving capabilities, a demand for high-resolution four-dimensional (4D) sensing has emerged(33). Such advanced sensing is essential not only for speed and range determination but also for accurately estimating targets’ azimuth and elevation with high resolution.

However, these radar systems primarily measure speed and range, offering limited azimuth angular resolution. The challenge of enhancing angular resolution has led to the extensive use of multiple-input multiple-output (MIMO) radar technology. MIMO radars synthesize a large virtual array aperture, significantly improving angular resolution with a manageable number of transmit and receive antennas (14; 15; 34; 3).

Further, signal processing techniques have been explored to further improve the angular resolution beyond what is achievable through digital beamforming technique that is implemented via fast Fourier transform (FFT). Super-resolution direction of arrival (DOA) estimation algorithms, such as compressive sensing (CS) (7; 6; 5) and the iterative adaptive approach (IAA) (39; 27), represent significant strides in this direction. Yet, their computational demand presents a formidable barrier to real-time implementation, especially in the dynamic context of automotive scenarios. Figure 1 illustrates how antenna aperture and super-resolution algorithms influence the quality of range-azimuth (RA) heatmaps. The high-resolution RA heatmaps contain rich information of the objects, including their shapes, facilitating object detection and classification through deep neural networks (40).

Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (1)

The adoption of deep learning (DL) techniques for radar image enhancement has yielded significant advances within the realm of image super-resolution, as demonstrated in computer vision research (1; 11; 16). Applying these methods to enhance azimuth resolution in RA heatmaps presents a significant opportunity for substantial improvement. Nonetheless, few studies have focused on generating super-resolution RA heatmaps using raw radar signals by exploiting radar domain knowledge. Approaches that consider generating radar super-resolution RA heatmaps as straightforward image-to-image or volume-to-volume tasks often overlook the critical domain knowledge of radar signal processing. This oversight can lead to data-intensive solutions, rely on excessively large networks, or fail to deliver optimal performance and scalability which are key concerns in automotive applications where rapid inference and compact model size are essential for on-chip implementation.

Research in the domain of super-resolution RA heatmap generation for automotive radar remains limited, with the majority of studies relying on FFT-generated ground truths from larger antenna arrays. To our best knowledge, no existing methods leverage RA heatmaps produced through super-resolution algorithms as ground truths. Moreover, these methods typically focus on smaller antenna arrays and do not explore the potential of varied training data sizes. This paper aims to close these gaps by introducing the Super-Resolution Angular Spectra Estimation Network (SR-SPECNet). Designed with radar signal processing expertise, SR-SPECNet advances super-resolution angular spectra generation by transforming RA heatmap enhancement into a manageable 1D azimuth super-resolution challenge. This transformation is supported by our novel data normalization approach and an signal-to-noise ratio (SNR)-guided loss function. SR-SPECNet is thoroughly evaluated across varied antenna apertures and training dataset sizes, a first in this research area, using a dedicated real-world dataset. Our experimental analysis confirms that SR-SPECNet achieves exceptional parameter efficiency, superior performance in imaging quality, and outstanding scalability. It consistently surpasses established benchmarks, showcasing its capability to adapt to various antenna configurations and dataset sizes.

The key contributions of our work include:

  • We introduced SR-SPECNet, a network designed for efficiency and effectiveness, capable of using single-snapshot measurement to robustly produce high-resolution automotive RA imaging typically obtained through IAA, but without IAA’s computational expense.

  • We adopted radar signal processing domain knowledge to guide the neural network design by translating the RA imaging as a 1D spectra estimation problem, introducing a novel real radar data normalization method and a SNR-guided loss function.

  • SR-SPECNet is the first network proven to robustly create high-resolution RA imaging with fast inference time from real automotive radar data featuring dynamic objects, demonstrating SR-SPECNet’s scalability, efficiency, and robust performance.

2 Related Work

The quest for enhanced radar imaging has predominantly focused on improving the azimuth resolution with limited number of antenna elements, given that range resolution can be augmented by increasing the bandwidth. In the automotive radar domain, digital beamforming (DBF) has emerged as the predominant DOA estimation algorithm, favored for its computational efficiency and robustness. This technique, typically implemented via FFT, however, faces limitations in angular resolution due to the Rayleigh criterion and is characterized by relatively high sidelobes (34; 26). Automotive radars, operating within highly dynamic environments, often have access to only a limited number of snapshots, sometimes as few as a single snapshot.This scenario renders super-resolution methods, such as Capon beamforming, MUSIC (30), and ESPRIT (29), which require multiple snapshots for covariance matrix estimation, less viable.

Compressive sensing (CS) techniques, which leverage the sparsity of target distributions in the angular domain, are highly effective in super-resolution, especially in snapshot-constrained settings (7; 5). Despite their potential, CS methods demand a dictionary matrix with low mutual coherence, which can be limiting. Alternatively, the Iterative Adaptive Approach (IAA) offers robust Direction of Arrival (DOA) estimation with limited snapshots, utilizing a nonparametric, iterative process (39; 27). IAA is advantageous for high-resolution radar imaging, surpassing subspace methods like MUSIC and ESPRIT and CS-based methods that falter under snapshot constraints or produce sparse results. However, IAA is computationally intensive, requiring large-scale matrix inversions at each step.Although fast and super-fast IAA variants (12; 38; 13) aim to reduce these demands by replacing matrix inversions with factorization, their benefits are marginal for small arrays and more pronounced for larger arrays, though computational challenges persist.

Recently, deep learning techniques have been applied to address the intricacies of azimuth super-resolution in RA maps.An adversarial network was tailored for super-resolution in micro-Doppler imagery (1), showcasing the potential of generative adversarial networks (GANs) in radar image enhancement. A U-Net architecture was employed for the super-resolution of weather radar maps (11), demonstrating the adaptability of deep convolutional networks to various radar data modalities. Notably, (16) ventured into extrapolating received antenna signals through a compact network, followed by the application of a 3D U-Net on the range-Doppler-azimuth data cube, facilitating the generation of super-resolution RA heatmaps.However, the existing body of work primarily leverages 2D or 3D network architectures, predicated on the assumption that the problem space necessitates multi-dimensional data processing to achieve enhanced resolution. This perspective, while valid, overlooks the potential efficiencies and novel insights that can be garnered from reinterpreting the challenge through a one-dimensional lens. To our best knowledge, no prior work has endeavored to address radar azimuth super-resolution within RA heatmaps using a 1D approach.In this paper, we close this gap by designing an efficient and effective deep neural network to achieving super-resolution radar imaging, through leveraging radar signal processing domain knowledge.

3 Radar Datasets

Dataset# of FramesData TypeResolutionRadar/Technology
nuScenes (4)40,0004000040,00040 , 000Sparse PCLowContinental ARS408
Oxford Radar (2)240,000240000240,000240 , 000RAHighNavtech Spinning Radar
RADIATE (32)44,0004400044,00044 , 000RAHighNavtech Spinning Radar
CRUW (37)396,241396241396,241396 , 241RALowTI AWR1843
Zendar (18)94,4609446094,46094 , 460ADC,RD,PCHighSAR
CARRADA (20)12,6661266612,66612 , 666RA,RD,RADLowTI AWR1843
RadarScenes (31)975975975975Dense PCHigh77GHz Middle-Range Radar
RADIal(25)25,0002500025,00025 , 000ADC,RD,PCHighValeo Middle Range DDM
View-of-Delft(22)8,69386938,6938 , 693PC+DopplerHighZF FRGen21 Radar
K-Radar (21)35,0003500035,00035 , 0004D TensorHighKAIST-Radar
Radatron (17)152,000152000152,000152 , 0004D TensorHighTI Cascade Imaging Radar
Ours17,0001700017,00017 , 000ADCHighTI Cascade Imaging Radar

Radar datasets for autonomous driving such as nuScenes (4), Oxford Radar RobotCar (2), RADIATE (32), and others, are summarized in Table 1.Technologies like spinning radar, utilized in the RADIATE and Oxford Radar RobotCar datasets, provide high-resolution 360-degree field-of-view (FOV) imagery, albeit at limited frame rates, which can introduce motion blur challenges. Dataset such as CARRADA employs single-chip Texas Instruments (TI) radar systems, offering modest angular resolutions exceeding 10superscript1010^{\circ}10 start_POSTSUPERSCRIPT ∘ end_POSTSUPERSCRIPT.The Zendar dataset, leveraging synthetic aperture radar (SAR) technology, excels in imaging static targets by integrating measurements from different vehicle positions.The View-of-Delft dataset takes advantage of the ZF FRGen21 radar’s long-range and high-resolution imaging capabilities, offering point cloud data with object annotations confined to a 50 meter range.

3.1 Our Dataset

Our approach demands detailed radar configuration parameters and intensive raw analog-to-digital converter (ADC) data processing to integrate super-resolution algorithms effectively and assess network performance across various antenna apertures. Hence, we created our own dataset by driving a Lexus RX450h SUV equipped with multi-modal sensors, including a TI imaging radar, Teledyne FLIR Blackfly S stereo cameras, and a Velodyne Ultra Puck VLP-32C LiDAR sensor, along urban streets, highways and campus roads. The centerpiece of our dataset is the TI cascaded imaging radar system (35), configured for MIMO operations with an array of 12 transmit (TX) and 16 receive (RX) antennas.The operational 9 TX and 16 RX antennas were arranged to form a virtual uniform linear array (ULA) of 86 elements, with half-wavelength spacing, rendering an azimuth resolution of roughly 1.2 degrees via FFT.Our dataset showcases the exceptional high-resolution capabilities of our radar configuration, as illustrated in Figure 2.

Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (2)

4 Method

We aim to transform raw ADC data into high-resolution RA maps. Unlike recent approaches that derive high-resolution ground truth from RA maps using an expanded antenna array (16), our method relies on RA maps generated with the same number of antennas but refined using IAA algorithms as our benchmark.

Figure 3 depicts our processing workflow. The input ADC data, IADCNfast×Nslow×Nchsubscript𝐼ADCsuperscriptsubscript𝑁fastsubscript𝑁slowsubscript𝑁chI_{\rm ADC}\in\mathbb{C}^{N_{\rm fast}\times N_{\rm slow}\times N_{\rm ch}}italic_I start_POSTSUBSCRIPT roman_ADC end_POSTSUBSCRIPT ∈ blackboard_C start_POSTSUPERSCRIPT italic_N start_POSTSUBSCRIPT roman_fast end_POSTSUBSCRIPT × italic_N start_POSTSUBSCRIPT roman_slow end_POSTSUBSCRIPT × italic_N start_POSTSUBSCRIPT roman_ch end_POSTSUBSCRIPT end_POSTSUPERSCRIPT, encapsulates three dimensions: Nfastsubscript𝑁fastN_{\rm fast}italic_N start_POSTSUBSCRIPT roman_fast end_POSTSUBSCRIPT for fast time samples, Nslowsubscript𝑁slowN_{\rm slow}italic_N start_POSTSUBSCRIPT roman_slow end_POSTSUBSCRIPT for slow time samples (or chirps), and Nchsubscript𝑁chN_{\rm ch}italic_N start_POSTSUBSCRIPT roman_ch end_POSTSUBSCRIPT for channels (or receivers). Through a 2D FFT to IADCsubscript𝐼ADCI_{\rm ADC}italic_I start_POSTSUBSCRIPT roman_ADC end_POSTSUBSCRIPT across both fast and slow time dimensions, we obtain range-Doppler-channel data, IRDCNRange×NDoppler×Nchsubscript𝐼RDCsuperscriptsubscript𝑁Rangesubscript𝑁Dopplersubscript𝑁chI_{\rm RDC}\in\mathbb{C}^{N_{\rm Range}\times N_{\rm Doppler}\times N_{\rm ch}}italic_I start_POSTSUBSCRIPT roman_RDC end_POSTSUBSCRIPT ∈ blackboard_C start_POSTSUPERSCRIPT italic_N start_POSTSUBSCRIPT roman_Range end_POSTSUBSCRIPT × italic_N start_POSTSUBSCRIPT roman_Doppler end_POSTSUBSCRIPT × italic_N start_POSTSUBSCRIPT roman_ch end_POSTSUBSCRIPT end_POSTSUPERSCRIPT. Subsequently, beam vectors y1×1×Nchysuperscript11subscript𝑁ch\textbf{y}\in\mathbb{C}^{1\times 1\times N_{\rm ch}}y ∈ blackboard_C start_POSTSUPERSCRIPT 1 × 1 × italic_N start_POSTSUBSCRIPT roman_ch end_POSTSUBSCRIPT end_POSTSUPERSCRIPT are extracted from each range-Doppler bin. These vectors are processed by SR-SPECNet to generate a super-resolution spectrum.This operation, performed on all beam vectors across all range-Doppler bins, yields the range-Doppler-azimuth data, IRDANRange×NDoppler×NAzimuthsubscript𝐼RDAsuperscriptsubscript𝑁Rangesubscript𝑁Dopplersubscript𝑁AzimuthI_{\rm RDA}\in\mathbb{R}^{N_{\rm Range}\times N_{\rm Doppler}\times N_{\rmAzimuth}}italic_I start_POSTSUBSCRIPT roman_RDA end_POSTSUBSCRIPT ∈ blackboard_R start_POSTSUPERSCRIPT italic_N start_POSTSUBSCRIPT roman_Range end_POSTSUBSCRIPT × italic_N start_POSTSUBSCRIPT roman_Doppler end_POSTSUBSCRIPT × italic_N start_POSTSUBSCRIPT roman_Azimuth end_POSTSUBSCRIPT end_POSTSUPERSCRIPT. Notably, this procedure is highly parallelizable, treating the dataset as a 2D matrix with a batch size of NRange×NDopplersubscript𝑁Rangesubscript𝑁Doppler{N_{\rm Range}\times N_{\rm Doppler}}italic_N start_POSTSUBSCRIPT roman_Range end_POSTSUBSCRIPT × italic_N start_POSTSUBSCRIPT roman_Doppler end_POSTSUBSCRIPT. The final high-resolution RA maps, MNRange×NAzimuth𝑀superscriptsubscript𝑁Rangesubscript𝑁AzimuthM\in\mathbb{R}^{N_{\rm Range}\times N_{\rm Azimuth}}italic_M ∈ blackboard_R start_POSTSUPERSCRIPT italic_N start_POSTSUBSCRIPT roman_Range end_POSTSUBSCRIPT × italic_N start_POSTSUBSCRIPT roman_Azimuth end_POSTSUBSCRIPT end_POSTSUPERSCRIPT, are achieved by averaging IRDAsubscript𝐼RDAI_{\rm RDA}italic_I start_POSTSUBSCRIPT roman_RDA end_POSTSUBSCRIPT over Doppler dimension.

Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (3)

To provide a clearer understanding of the beam vector in the context of automotive radar signal processing, its signal model can be articulated as follows:

𝐲=A(θ)s+n,𝐲A𝜃sn\displaystyle{\bf y}=\textbf{A}(\theta)\textbf{s}+\textbf{n},bold_y = A ( italic_θ ) s + n ,(1)

where θ𝜃\thetaitalic_θ encapsulates the DOA of targets, 𝐧𝐧\bf{n}bold_n signifies a complex Nch×1subscript𝑁𝑐1N_{ch}\times 1italic_N start_POSTSUBSCRIPT italic_c italic_h end_POSTSUBSCRIPT × 1 white Gaussian noise vector, and 𝐀(θ)=[𝐚(θ1),𝐚(θ2),,𝐚(θK)]𝐀𝜃𝐚subscript𝜃1𝐚subscript𝜃2𝐚subscript𝜃𝐾{\bf A}(\theta)=\left[{\bf a}(\theta_{1}),{\bf a}(\theta_{2}),\cdots,{\bf a}(%\theta_{K})\right]bold_A ( italic_θ ) = [ bold_a ( italic_θ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT ) , bold_a ( italic_θ start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT ) , ⋯ , bold_a ( italic_θ start_POSTSUBSCRIPT italic_K end_POSTSUBSCRIPT ) ] represents the Nch×Ksubscript𝑁𝑐𝐾N_{ch}\times Kitalic_N start_POSTSUBSCRIPT italic_c italic_h end_POSTSUBSCRIPT × italic_K array manifold matrix for K𝐾Kitalic_K targets. The array response vector a(θ)a𝜃\textbf{a}(\theta)a ( italic_θ ) is given as a(θ)=[1,e2πd2λsinθ,,e2πdNchλsinθ]Ta𝜃superscript1superscript𝑒2𝜋subscript𝑑2𝜆𝜃superscript𝑒2𝜋subscript𝑑subscript𝑁𝑐𝜆𝜃𝑇{\textbf{a}}(\theta)=\left[1,e^{\frac{2\pi d_{2}}{\lambda}\sin{\theta}},\cdots%,e^{\frac{2\pi d_{N_{ch}}}{\lambda}\sin{\theta}}\right]^{T}a ( italic_θ ) = [ 1 , italic_e start_POSTSUPERSCRIPT divide start_ARG 2 italic_π italic_d start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT end_ARG start_ARG italic_λ end_ARG roman_sin italic_θ end_POSTSUPERSCRIPT , ⋯ , italic_e start_POSTSUPERSCRIPT divide start_ARG 2 italic_π italic_d start_POSTSUBSCRIPT italic_N start_POSTSUBSCRIPT italic_c italic_h end_POSTSUBSCRIPT end_POSTSUBSCRIPT end_ARG start_ARG italic_λ end_ARG roman_sin italic_θ end_POSTSUPERSCRIPT ] start_POSTSUPERSCRIPT italic_T end_POSTSUPERSCRIPT. In this model, dnsubscript𝑑𝑛d_{n}italic_d start_POSTSUBSCRIPT italic_n end_POSTSUBSCRIPT denotes the spacing between the n𝑛nitalic_n-th element and the reference element, and 𝐬=[s1,s2,,sK]T𝐬superscriptsubscript𝑠1subscript𝑠2subscript𝑠𝐾𝑇{\bf s}=[s_{1},s_{2},\cdots,s_{K}]^{T}bold_s = [ italic_s start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , italic_s start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT , ⋯ , italic_s start_POSTSUBSCRIPT italic_K end_POSTSUBSCRIPT ] start_POSTSUPERSCRIPT italic_T end_POSTSUPERSCRIPT represents the vector of source strengths.

4.1 SR-SPECNet for Azimuth Super-Resolution

CS and IAA stand out as widely recognized super-resolution DOA estimation algorithms tailored for single-snapshot radar data. In contrast to CS, which yields spectra consisting solely of discrete points, the IAA generates continuous spectra that include estimated reflection coefficients. This particular feature renders IAA more apt for the generation of RA maps.

As a super-resolution DOA estimation algorithm working under single-snapshot, IAA generates continuous spectra that include estimated reflection coefficients, rendering IAA more apt for the generation of RA maps.

4.1.1 Iterative Adaptive Approach (IAA)

IAA is a data-dependent, nonparametric algorithm (39). By discretizing the DOA space into an L𝐿Litalic_L point grid, the array manifold is defined as 𝐀(θ)=[𝐚(θ1),,𝐚(θL)]𝐀𝜃𝐚subscript𝜃1𝐚subscript𝜃𝐿\mathbf{A}(\theta)=\left[\mathbf{a}(\theta_{1}),\cdots,\mathbf{a}(\theta_{L})\right]bold_A ( italic_θ ) = [ bold_a ( italic_θ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT ) , ⋯ , bold_a ( italic_θ start_POSTSUBSCRIPT italic_L end_POSTSUBSCRIPT ) ] with a(θ)a𝜃{\textbf{a}}(\theta)a ( italic_θ ) being the array steering vector. The fictitious covariance matrix of 𝐲𝐲\bf{y}bold_y is represented as 𝐑f=𝐀(θ)𝐏𝐀H(θ)subscript𝐑𝑓𝐀𝜃superscript𝐏𝐀𝐻𝜃{\bf R}_{f}={\bf A}(\theta){\bf P}{\bf A}^{H}(\theta)bold_R start_POSTSUBSCRIPT italic_f end_POSTSUBSCRIPT = bold_A ( italic_θ ) bold_PA start_POSTSUPERSCRIPT italic_H end_POSTSUPERSCRIPT ( italic_θ ), where P𝑃Pitalic_P is a L×L𝐿𝐿L\times Litalic_L × italic_L diagonal matrix with the l𝑙litalic_l-th diagonal element being Pl=|s^l|2subscript𝑃𝑙superscriptsubscript^𝑠𝑙2P_{l}=|{\hat{s}}_{l}|^{2}italic_P start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT = | over^ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT | start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT, and s^lsubscript^𝑠𝑙{\hat{s}}_{l}over^ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT is the source reflection coefficient corresponding to direction θlsubscript𝜃𝑙\theta_{l}italic_θ start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT. IAA iteratively estimates the reflection coefficient s^^𝑠\hat{s}over^ start_ARG italic_s end_ARG and updates the fictitious covariance matrix by minimizing a weighted least-square (WLS) cost function 𝐲sl𝐚(θl)𝐐1(θl)2superscriptsubscriptdelimited-∥∥𝐲subscript𝑠𝑙𝐚subscript𝜃𝑙superscript𝐐1subscript𝜃𝑙2\lVert{\bf y}-s_{l}{\bf a}(\theta_{l})\rVert_{{\bf Q}^{-1}(\theta_{l})}^{2}∥ bold_y - italic_s start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT bold_a ( italic_θ start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT ) ∥ start_POSTSUBSCRIPT bold_Q start_POSTSUPERSCRIPT - 1 end_POSTSUPERSCRIPT ( italic_θ start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT ) end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT, where 𝐗𝐐1(θl)2=Δ𝐗H𝐐1(θl)𝐗superscriptsubscriptdelimited-∥∥𝐗superscript𝐐1subscript𝜃𝑙2Δsuperscript𝐗𝐻superscript𝐐1subscript𝜃𝑙𝐗\lVert{\bf X}\rVert_{{\bf Q}^{-1}(\theta_{l})}^{2}\overset{\Delta}{=}{\bf X}^{%H}{\bf Q}^{-1}(\theta_{l}){\bf X}∥ bold_X ∥ start_POSTSUBSCRIPT bold_Q start_POSTSUPERSCRIPT - 1 end_POSTSUPERSCRIPT ( italic_θ start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT ) end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT overroman_Δ start_ARG = end_ARG bold_X start_POSTSUPERSCRIPT italic_H end_POSTSUPERSCRIPT bold_Q start_POSTSUPERSCRIPT - 1 end_POSTSUPERSCRIPT ( italic_θ start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT ) bold_X and 𝐐(θl)=𝐑fPl𝐚(θl)𝐚H(θl)𝐐subscript𝜃𝑙subscript𝐑𝑓subscript𝑃𝑙𝐚subscript𝜃𝑙superscript𝐚𝐻subscript𝜃𝑙{\bf Q}(\theta_{l})={\bf R}_{f}-P_{l}{\bf a}(\theta_{l}){\bf a}^{H}(\theta_{l})bold_Q ( italic_θ start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT ) = bold_R start_POSTSUBSCRIPT italic_f end_POSTSUBSCRIPT - italic_P start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT bold_a ( italic_θ start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT ) bold_a start_POSTSUPERSCRIPT italic_H end_POSTSUPERSCRIPT ( italic_θ start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT ).

4.1.2 SR-SPECNet

SR-SPECNet’s primary objective is to transform input beam vectors into super-resolution IAA spectra. The IAA, functioning as an advanced beamforming algorithm, iteratively refines a reconstructed covariance matrix to estimate the spectrum as 𝐬^IAA=𝐖H𝐲subscript^𝐬IAAsuperscript𝐖𝐻𝐲\hat{\mathbf{s}}_{\rm IAA}=\mathbf{W}^{H}\mathbf{y}over^ start_ARG bold_s end_ARG start_POSTSUBSCRIPT roman_IAA end_POSTSUBSCRIPT = bold_W start_POSTSUPERSCRIPT italic_H end_POSTSUPERSCRIPT bold_y, where 𝐖Nch×L𝐖superscriptsubscript𝑁ch𝐿\mathbf{W}\in\mathbb{C}^{N_{\rm ch}\times L}bold_W ∈ blackboard_C start_POSTSUPERSCRIPT italic_N start_POSTSUBSCRIPT roman_ch end_POSTSUBSCRIPT × italic_L end_POSTSUPERSCRIPT are the beamforming weights. At each beamforming angle θLsubscript𝜃𝐿\theta_{L}italic_θ start_POSTSUBSCRIPT italic_L end_POSTSUBSCRIPT, the array’s response, s^l=𝐖H(θl)𝐲subscript^𝑠𝑙superscript𝐖𝐻subscript𝜃𝑙𝐲\hat{s}_{l}=\mathbf{W}^{H}(\theta_{l})\mathbf{y}over^ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT = bold_W start_POSTSUPERSCRIPT italic_H end_POSTSUPERSCRIPT ( italic_θ start_POSTSUBSCRIPT italic_l end_POSTSUBSCRIPT ) bold_y, parallels the output process of a multi-layer perceptron (MLP), underscoring the suitability of using an MLP for this application (19).

Designed as a four-layer MLP, SR-SPECNet mirrors the mathematical operations in the IAA algorithm. It processes the input signal𝐲Nch𝐲superscriptsubscript𝑁ch{\bf y}\in{\mathbb{C}}^{N_{\rm ch}}bold_y ∈ blackboard_C start_POSTSUPERSCRIPT italic_N start_POSTSUBSCRIPT roman_ch end_POSTSUBSCRIPT end_POSTSUPERSCRIPT, by separating its real and imaginary parts and concatenating them into a real-valued input𝐲¯2Nch¯𝐲superscript2subscript𝑁ch{\bf\bar{y}}\in\mathbb{R}^{2*N_{\rm ch}}over¯ start_ARG bold_y end_ARG ∈ blackboard_R start_POSTSUPERSCRIPT 2 ∗ italic_N start_POSTSUBSCRIPT roman_ch end_POSTSUBSCRIPT end_POSTSUPERSCRIPT. This approach ensures the preservation of crucial phase information, as complex value multiplication inherently involves both the real and imaginary parts.

4.2 Data Preprocessing

Proper data normalization is crucial for training neural networks, especially for regression tasks.Different from simulated signals with controlled factors like SNR, target reflection, and target number, real-world signals add unpredictability in SNR and reflections, challenging normalization. SNR varies significantly within a radar frame and from frame to frame. Maintaining a comparable intensity among beam vectors is crucial for constructing accurate RA heatmaps, where factors like SNR, target reflection intensity, and number of targets are precisely controlled. Real-world signals, however, introduce complexities not present in simulated environments: SNR and target reflection intensities vary unpredictably, complicating the task of normalizing inputs and labels. Within a single radar frame, the range of SNR values across beam vectors can be vast, with some vectors lacking targets altogether or presenting very low SNR. Adding to the complexity is the necessity to maintain the relative intensity among beam vectors within each radar frame, a critical aspect for accurately constructing the final RA heatmap.

To overcome the normalization challenges posed by the variability in real-world signals,we introduce a frequency domain normalization method designed to produce consistent and interpretable inputs for neural network training. This approach entails determining a normalization factor, α𝛼\alphaitalic_α, for each beam vector, calculated as the maximum absolute value of the frequency spectra, obtained by multiplying 𝐀Hsuperscript𝐀𝐻\mathbf{A}^{H}bold_A start_POSTSUPERSCRIPT italic_H end_POSTSUPERSCRIPT to the beam vector, equivalent to an FFT operation, and then divided by the total number of elements, Nchsubscript𝑁chN_{\rm ch}italic_N start_POSTSUBSCRIPT roman_ch end_POSTSUBSCRIPT:

α=max(|𝐀H𝐲Nch|).𝛼superscript𝐀𝐻𝐲subscript𝑁ch\displaystyle\alpha=\max\left(\left|\frac{\mathbf{A}^{H}\mathbf{y}}{N_{\rm ch}%}\right|\right).italic_α = roman_max ( | divide start_ARG bold_A start_POSTSUPERSCRIPT italic_H end_POSTSUPERSCRIPT bold_y end_ARG start_ARG italic_N start_POSTSUBSCRIPT roman_ch end_POSTSUBSCRIPT end_ARG | ) .(2)

Subsequently, the raw signal 𝐲𝐲\mathbf{y}bold_y is normalized using α𝛼\alphaitalic_α to yield 𝐲norm=𝐲/αsubscript𝐲norm𝐲𝛼\mathbf{y}_{\rm norm}={\mathbf{y}}/{\alpha}bold_y start_POSTSUBSCRIPT roman_norm end_POSTSUBSCRIPT = bold_y / italic_α, ensuring that the signal levels are stable across varying SNR conditions. Similarly, the label, represented by the IAA spectra 𝐬^IAAsubscript^𝐬IAA\hat{\mathbf{s}}_{\rm IAA}over^ start_ARG bold_s end_ARG start_POSTSUBSCRIPT roman_IAA end_POSTSUBSCRIPT, is normalized to 𝐬norm=𝐬^IAA/αsubscript𝐬normsubscript^𝐬IAA𝛼\mathbf{s}_{\rm norm}={\hat{\mathbf{s}}_{\rm IAA}}/{\alpha}bold_s start_POSTSUBSCRIPT roman_norm end_POSTSUBSCRIPT = over^ start_ARG bold_s end_ARG start_POSTSUBSCRIPT roman_IAA end_POSTSUBSCRIPT / italic_α. This normalization strategy effectively scales the signal and the IAA spectra so that their values fall within a comparable range, thereby facilitating more effective network training.Further, α𝛼\alphaitalic_α maintains a relative intensity among all beam vectors within each radar frame, ensuring that the spatial relationships that are critical for accurate RA heatmap synthesis are preserved.Moreover, α𝛼\alphaitalic_α preserves the relative intensity across all beam vectors in a radar frame, maintaining the spatial relationships essential for accurate synthesis of RA heatmaps. It’s important to note that this normalization process is exclusively needed during training and is not required for generating super-resolution RA heatmaps with test data, which stands as a significant advantage.This is attributed to the linear relationship between the beam vector and its corresponding spectra, allowing for direct processing without the need for normalization in the testing phase.

4.3 SNR-Guided Loss Function

The normalization factor α𝛼\alphaitalic_α, which represents the maximum value in the signal’s frequency domain, is directly proportional to the signal’s SNR. A higherα𝛼\alphaitalic_α suggests a higher SNR, positively influencing the quality of the final RA heatmap. We introduce an SNR-guided loss function similar to a weighted mean squared error (MSE), designed to prioritize higher SNR signals during training. The loss function is defined as:

SNR=α1Li=1L(sis^i)2,subscriptSNR𝛼1𝐿superscriptsubscript𝑖1𝐿superscriptsubscript𝑠𝑖subscript^𝑠𝑖2\displaystyle\mathcal{L}_{\rm SNR}=\alpha\cdot\frac{1}{L}\sum_{i=1}^{L}(s_{i}-%\hat{s}_{i})^{2},caligraphic_L start_POSTSUBSCRIPT roman_SNR end_POSTSUBSCRIPT = italic_α ⋅ divide start_ARG 1 end_ARG start_ARG italic_L end_ARG ∑ start_POSTSUBSCRIPT italic_i = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_L end_POSTSUPERSCRIPT ( italic_s start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT - over^ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ) start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT ,

where L𝐿Litalic_L is the number of angle grid points of the spectra, sisubscript𝑠𝑖s_{i}italic_s start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT and s^isubscript^𝑠𝑖\hat{s}_{i}over^ start_ARG italic_s end_ARG start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT are the actual and predicted values at θisubscript𝜃𝑖\theta_{i}italic_θ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT. This approach ensures that our model is finely tuned to emphasize higher quality signals.

5 Experiment

We train and evaluate our SR-SPECNet model using our own dataset, which comprises 17,000 frames of raw ADC radar data. To promote data diversity and minimize the redundancy of consecutive frames, we strategically selected every tenth frame from the dataset, yielding 1,700 frames, with the initial 1,400 frames dedicated to training the model, and the subsequent 300 frames reserved for testing. We intentionally structured the training frames into three subsets to simulate real-world data collection scenarios of limited time periods: a ‘small’ dataset with the initial 200 frames (akin to a 200-second data collection period), a ‘medium’ dataset comprising the first 700 frames, and a ‘large’ dataset that includes all 1,400 frames. This segmentation aims to test our model’s performance and adaptability under varying lengths of data availability.

5.1 Benchmarks

To evaluate SR-SPECNet’s effectiveness, we compare it with models designed to enhance spatial resolution. This comparison includes a 2D U-Net (28), which transforms low-resolution RA heatmaps into high-resolution equivalents, and the RAD-UNet (16), referred to as a 3D U-Net, that upgrades low-resolution range-azimuth-Doppler (RAD) data to high-resolution RAD imagery. We exclude pixel-based super-resolution networks like SRGAN, which increase resolution by adding pixels. These models do not meet the specific requirements of radar imaging, where resolution is not directly related to pixel count (16).

5.2 Evaluation Metrics

The RA map is a grayscale image, normalized between 0 to 1, for both generated and ground truth images. To comprehensively evaluate the quality of high-resolution RA heatmaps, we use established image evaluation metrics. PSNR, measured in dB, and SSIM, ranging from 0 to 1, assess image quality where higher values indicate better quality. NMSE also ranges from 0 to 1 and quantifies prediction accuracy by comparing the mean squared error to the variance of actual values, with lower values indicating more accurate predictions. Together, PSNR, SSIM, and NMSE provide a robust framework for assessing image fidelity, error magnitude, and compositional changes affecting perceived quality.

5.3 Implementation Details

Our radar configuration is characterized by fast-time samples Nfast=256subscript𝑁fast256N_{\rm fast}=256italic_N start_POSTSUBSCRIPT roman_fast end_POSTSUBSCRIPT = 256, slow-time samples Nslow=64subscript𝑁slow64N_{\rm slow}=64italic_N start_POSTSUBSCRIPT roman_slow end_POSTSUBSCRIPT = 64, and post-MIMO processing, resulting in beam vectors, each with 86868686 elements.We truncated range of the radar data cube, IRDCsubscript𝐼RDCI_{\rm RDC}italic_I start_POSTSUBSCRIPT roman_RDC end_POSTSUBSCRIPT, to keep the first 100100100100 elements along the range axis, resulting in a truncated dataset IRDCtrunc100×64×Nchsuperscriptsubscript𝐼RDCtruncsuperscript10064subscript𝑁chI_{\rm RDC}^{\rm trunc}\in\mathbb{C}^{100\times 64\times N_{\rm ch}}italic_I start_POSTSUBSCRIPT roman_RDC end_POSTSUBSCRIPT start_POSTSUPERSCRIPT roman_trunc end_POSTSUPERSCRIPT ∈ blackboard_C start_POSTSUPERSCRIPT 100 × 64 × italic_N start_POSTSUBSCRIPT roman_ch end_POSTSUBSCRIPT end_POSTSUPERSCRIPT. This truncation strategy is informed by the observation that significant target information is concentrated within the first 50505050 meters of the collected data.

We embark on an exploratory analysis of the effect of antenna aperture size on network performance. To this end, we selected a 10-element antenna array to represent a smaller aperture and a 40-element array for a larger aperture. The choice of a 40-element array as the larger aperture is strategically made, considering that it provides sufficient resolution, achieving approximately 1superscript11^{\circ}1 start_POSTSUPERSCRIPT ∘ end_POSTSUPERSCRIPT angular resolution using the IAA algorithm.

We set the angular grid size to L=256𝐿256L=256italic_L = 256 for frequency domain uniformity. The labels for our 10-element and 40-element antenna arrays stem from their IAA spectra, ensuring our network’s performance evaluation remains consistent across varying apertures. SR-SPECNet comprises four fully connected layers, with the first three followed by ReLU activation functions and output sizes of 2048, 1024, and 512, respectively. The final layer’s output size matches L𝐿Litalic_L. As input signals are complex, we concatenate their real and imaginary parts into a real-valued vector, serving as the input to SR-SPECNet.

We implemented SR-SPECNet and benchmark models using PyTorch, standardizing training with the Adam optimizer at a learning rate of 0.0001 for 500 epochs. Training was accelerated on four Nvidia RTX A6000 GPUs for efficiency.

5.4 High-Resolution RA Heatmap

We study the performance of deep neural networks in generating high-resolution RA heatmaps. In pursuit of this, we evaluated SR-SPECNet and SR-SPECNet+, which were trained with MSE loss and our SNR-guided loss, respectively, against established benchmark models. As delineated in Tables 2 and 3, our models were tested using 10 and 40 antenna elements across small, medium, and large dataset sizes.

For the 10-element antenna configuration, Table 2 highlights the strength of our methodology. SR-SPECNet+ emerges as the leading model, eclipsing both the 2D and 3D U-Net benchmarks as well as SR-SPECNet across all dataset sizes. Its edge in performance can be attributed to the incorporation of domain knowledge through the use of an SNR-guided loss function, which fine-tunes the training process to emphasize data with higher signal integrity. This advanced loss function allows SR-SPECNet+ to achieve the lowest NMSE and the highest SSIM and PSNR scores, demonstrating its effectiveness even in small dataset scenarios. SR-SPECNet itself outpaces conventional U-Net models, underscoring the value of generating RA heatmap through a 1D super-resolution lens. The comparative results make it evident that the domain-specific enhancements in SR-SPECNet+ significantly boost its ability to generate high-quality heatmaps, particularly when the number of antenna elements is limited.

Modelssmallmediumlarge
NMSE\downarrowSSIM\uparrowPSNR\uparrowNMSE\downarrowSSIM\uparrowPSNR\uparrowNMSE\downarrowSSIM\uparrowPSNR\uparrow
2D U-Net0.3210.78227.7800.2330.82029.1820.1040.87732.736
3D U-Net0.7630.84126.6550.2050.89431.3180.1320.91733.958
SR-SPECNet0.1680.90930.6830.0920.94633.4460.0770.95534.326
SR-SPECNet+0.0800.95034.0060.0630.96235.3500.0560.96535.884

Shifting the focus to the 40-element antenna configuration, as presented in Table 3, SR-SPECNet+ maintains its exceptional standard. It substantially betters its NMSE and improves SSIM and PSNR values, asserting its robustness across all data volumes. SR-SPECNet retains a strong performance profile, lending weight to the concept that a 1D super-resolution approach effectively facilitates the production of high-quality heatmaps without extensive training data. This evidence underscores our model’s proficiency in handling data from varying antenna aperture sizes and confirms the strategic value of the 1D super-resolution methodology, which maximizes training data efficiency and is conducive to high-resolution radar imaging applications.

Modelssmallmediumlarge
NMSE\downarrowSSIM\uparrowPSNR\uparrowNMSE\downarrowSSIM\uparrowPSNR\uparrowNMSE\downarrowSSIM\uparrowPSNR\uparrow
2D U-Net0.2330.90334.6760.1730.93436.2760.1240.94136.887
3D U-Net1.2550.81331.6491.3200.81633.1060.2920.94637.378
SR-SPECNet0.2680.92934.1350.1700.94236.0840.1150.96038.100
SR-SPECNet+0.1440.95136.9240.1220.95937.9160.0930.96539.066

5.5 Complexity and Scalability

The 2D and 3D U-Nets, designed for processing low-resolution RA and RAD heatmaps, have fixed numbers of trainable parameters for different size antenna arrays: approximately 31.0M for the 2D U-Net and 51.8M for the 3D U-Net. Their inference times are 6.14 ms and 6.98 ms, respectively, indicating the increased computational demands of higher-dimensional data processing. In contrast, our SR-SPECNet achieves notable efficiency, operating with just 1M parameters and a swift inference time of 3.12 ms. This reduction in time and model size enhances the speed and resource efficiency of our method, ideal for real-time automotive radar applications on embedded CPUs. Meanwhile, the IAA’s inference times are 12.3 ms for 10-element vectors and 30.73 ms for 40-element vectors, underscoring its computational challenges.

Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (4)

Our SR-SPECNet model is designed for adaptability, accepting 1D beam vectors as input, which provides superior scalability across varying NRangesubscript𝑁RangeN_{\text{Range}}italic_N start_POSTSUBSCRIPT Range end_POSTSUBSCRIPT and NDopplersubscript𝑁DopplerN_{\text{Doppler}}italic_N start_POSTSUBSCRIPT Doppler end_POSTSUBSCRIPT values. As outlined in Section 5.3, our training dataset is configured with NRange=100subscript𝑁Range100N_{\text{Range}}=100italic_N start_POSTSUBSCRIPT Range end_POSTSUBSCRIPT = 100 and NDoppler=64subscript𝑁Doppler64N_{\text{Doppler}}=64italic_N start_POSTSUBSCRIPT Doppler end_POSTSUBSCRIPT = 64. However, as demonstrated in Figure4, SR-SPECNet effortlessly handles different sizes of these parameters. We generate low-resolution and ground truth heatmaps using FFT and the IAA algorithm, respectively. This remarkable scalability highlights the advantages of our approach, confirming its suitability for diverse and dynamic radar imaging scenarios.

5.6 Generalizability

Our training and test datasets feature a rich variety of signals, with each radar frame containing thousands of signals to ensure diversity. To further test the model’s generalizability, we evaluated our pre-trained SR-SPECNet on 10,500 radar frames from Radartron (17), which present different NRangesubscript𝑁RangeN_{\text{Range}}italic_N start_POSTSUBSCRIPT Range end_POSTSUBSCRIPT and NDopplersubscript𝑁DopplerN_{\text{Doppler}}italic_N start_POSTSUBSCRIPT Doppler end_POSTSUBSCRIPT. Unlike the 2D and 3D UNets, which could not be applied directly due to these variations, our 1D approach was seamlessly implemented, showcasing its superior scalability. Performance metrics include a NMSE of 0.025, SSIM of 0.983, and PSNR of 37.858 for the 10-element configuration, and NMSE of 0.0743, SSIM of 0.970, and PSNR of 42.941 for the 40-element configuration.

Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (5)

5.7 Visualization of RA maps

To evaluate the quality of high-resolution RA heatmap by SR-SPECNet and SR-SPECNet+, we visually compare the outputs against those from baseline models, i.e., 2D U-Net and 3D U-Net, as depicted in Fig. 5. Each model is trained on the ‘large’ dataset to obtain the best performance. Ground truth heatmaps generated with the IAA, have much better resolution than the LR images.Further, IAA suppresses sidelobes, yielding much clearer heatmaps. This underscores the benefits of using IAA-derived as ground truth for learning, rather than using FFT-generated heatmaps with a larger antenna aperture, which may not always be practically available due to hardware cost and the complexities involved in MIMO technology implementations.Fig. 5 demonstrates thatthe RA heatmaps that are respectively produced by SR-SPECNet and SR-SPECNet+ with a SNR-guided loss function, exhibit a noticeable improvement over the heatmaps generated by 2D and 3D U-Net, aligning with the quantitative metrics presented in Tables 2 and 3. These results collectively affirm the effectiveness of our proposed 1D methodologies over the 2D and 3D methods.

6 Conclusions

In this study, we have advanced automotive radar imaging by introducing SR-SPECNet, a novel 1D network leveraging IAA-generated RA heatmaps as ground truth and integrating a unique SNR-guided loss function for super-resolution RA heatmap generation. Our approach, emphasizing a 1D signal processing perspective, has demonstrated superior performance on real radar measurements in terms of automotive radar imaging quality, scalability, and efficiency across varying antenna configurations and dataset sizes. These contributions enhance the fidelity of radar imaging in autonomous vehicles. They also opens avenues for future research, especially with our commitment to sharing our own radar dataset and source code resources with the research community. This work underscores the potential of deep learning-enhanced radar processing in improving navigational safety and robust perception in autonomous vehicles.

7 Limitation & Future works

Our proposed method opens up possibilities for super-resolution imaging across four-dimensional (4D) radar parameters, range, Doppler, azimuth, and elevation. Extending our method to fully support 4D radar imaging represents a key direction for our future research. Currently, our network is optimized for ULA only; thus, adapting our model to accommodate arbitrary array geometries, including various sparse configurations and arrays affected by random antenna failures, is crucial. Additionally, we plan to enhance our network’s performance across diverse dynamic ranges and in scenarios characterized by extremely low SNR. These improvements are vital for advancing the robustness and applicability of our technology in real-world environments.

References

  • Armanious etal. [2019]K.Armanious, S.Abdulatif, F.Aziz, U.Schneider, and B.Yang.An adversarial super-resolution remedy for radar design trade-offs.In 2019 27th European Signal Processing Conference (EUSIPCO), pages 1–5, 2019.doi: 10.23919/EUSIPCO.2019.8902510.
  • Barnes etal. [2020]D.Barnes, M.Gadd, P.Murcutt, P.Newman, and I.Posner.The Oxford radar robotcar dataset: A radar extension to the oxford robotcar dataset.In Proc. IEEE International Conference on Robotics and Automation (ICRA), Paris, France, May 31-Oct. 31, 2020.
  • Bergin and Guerci [2018]J.Bergin and J.R. Guerci.MIMO Radar: Theory and Application.Boston, MA, Artech House, 2018.
  • Caesar etal. [2020]H.Caesar, V.Bankiti, A.H. Lang, S.Vora, V.E. Liong, Q.Xu, A.Krishnan, Y.Pan, G.Baldan, and O.Beijbom.nuScenes: A multimodal dataset for autonomous driving.IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 16-18, 2020.
  • Candes and Fernandez-Granda [2014]E.Candes and C.Fernandez-Granda.Towards a mathematical theory of super-resolution.Communications on Pure and Applied Mathematics, 67(6):906–956, 2014.
  • Candès and Romberg [2007]E.Candès and J.Romberg.Sparsity and incoherence in compressive sampling.Inverse problems, 23(3):969, 2007.
  • Donoho [2006]D.L. Donoho.Compressed sensing.IEEE Transactions on Information Theory, 52(4):1289–1306, 2006.
  • Duggal etal. [2020]G.Duggal, S.Vishwakarma, K.V. Mishra, and S.S. Ram.Doppler-resilient 802.11ad-based ultrashort range automotive joint radar-communications system.IEEE Transactions on Aerospace and Electronic Systems, 56(5):4035–4048, 2020.
  • Engels etal. [2017]F.Engels, P.Heidenreich, A.M. Zoubir, F.K. Jondral, and M.Wintermantel.Advances in automotive radar: A framework on computationally efficient high-resolution frequency estimation.IEEE Signal Processing Magazine, 34(2):36–46, 2017.
  • Engels etal. [2021]F.Engels, P.Heidenreich, M.Wintermantel, L.Stäcker, M.AlKadi, and A.M. Zoubir.Automotive radar signal processing: Research directions and practical challenges.IEEE Journal of Selected Topics in Signal Processing, 15(4):865–878, 2021.
  • Geiss and Hardin [2020]A.Geiss and J.C. Hardin.Radar super resolution using a deep convolutional neural network.Journal of Atmospheric and Oceanic Technology, 37(12):2197–2207, 2020.
  • Glentis and Jakobsson [2011]G.-O. Glentis and A.Jakobsson.Efficient implementation of iterative adaptive approach spectral estimation techniques.IEEE Transactions on Signal Processing, 59(9):4154–4167, 2011.doi: 10.1109/TSP.2011.2145376.
  • Glentis and Jakobsson [2012]G.O. Glentis and A.Jakobsson.Superfast approximative implementation of the IAA spectral estimate.IEEE Transactions on Signal Processing, 60(1):472–478, 2012.doi: 10.1109/TSP.2011.2170979.
  • Li and Stoica [2007]J.Li and P.Stoica.MIMO radar with colocated antennas.IEEE Signal Processing Magazine, 24(5):106–114, 2007.doi: 10.1109/MSP.2007.904812.
  • Li and Stoica [2009]J.Li and P.Stoica.MIMO Radar Signal Processing.Hoboken, NJ, Wiley, 2009.
  • Li etal. [2023]Y.-J. Li, S.Hunt, J.Park, M.O’Toole, and K.Kitani.Azimuth super-resolution for FMCW radar in autonomous driving.In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 17504–17513, 2023.
  • Madani etal. [2022]S.Madani, J.Guan, W.Ahmed, S.Gupta, and H.Hassanieh.Radatron: Accurate detection using multi-resolution cascaded MIMO radar.In European Conference on Computer Vision (ECCV), pages 160–178, 2022.
  • Mostajabi etal. [2020]M.Mostajabi, C.M. Wang, D.Ranjan, and G.Hsyu.High resolution radar dataset for semi-supervised learning of dynamic objects.In Proc. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Seattle, WA, June 14-19, 2020.
  • Naumovski and Carrasco [1995]M.Naumovski and R.Carrasco.Neural network beamformer for narrow-band HF transmission.In IEE Colloquium on HF Antennas and Propagation, pages 5/1–5/8, 1995.doi: 10.1049/ic:19951273.
  • Ouaknine etal. [2021]A.Ouaknine, A.Newson, J.Rebut, F.Tupin, and P.Perez.CARRADA dataset: Camera and automotive radar with range-angle-Doppler annotations.In Proc. 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, Jan. 10-15, 2021.
  • Paek etal. [2022]D.-H. Paek, S.-H. Kong, and K.T. Wijaya.K-radar: 4D radar object detection for autonomous driving in various weather conditions.In Thirty-sixth Conference on Neural Information Processing Systems Datasets and Benchmarks Track, 2022.URL https://openreview.net/forum?id=W_bsDmzwaZ7.
  • Palffy etal. [2022]A.Palffy, E.Pool, S.Baratam, J.Kooij, and D.Gavrila.Multi-class road user detection with 3+1D radar in the view-of-Delft dataset.IEEE Robotics and Automation Letters, 7(2):4961–4968, 2022.doi: 10.1109/LRA.2022.3147324.
  • Patole etal. [2017]S.Patole, M.Torlak, D.Wang, and M.Ali.Automotive radars: A review of signal processing techniques.IEEE Signal Processing Magazine, 34(2):22–35, 2017.
  • Peng etal. [2022]Z.Peng, C.Li, and F.Uysal.Modern Radar for Automotive Applications.London, UK: IET, 2022.
  • Rebut etal. [2022]J.Rebut, A.Ouaknine, W.Malik, and P.Pérez.Raw high-definition radar for multi-task learning.In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 17021–17030, 2022.
  • Richards [2022]M.A. Richards.Fundamentals of Radar Signal Processing, 3rd Ed.New York, McGraw-Hill, 2022.
  • Roberts etal. [2010]W.Roberts, P.Stoica, J.Li, T.Yardibi, and F.Sadjadi.Iterative adaptive approaches to MIMO radar imaging.4(1):5–20, 2010.
  • Ronneberger etal. [2015]O.Ronneberger, P.Fischer, and T.Brox.U-net: Convolutional networks for biomedical image segmentation.In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18, pages 234–241. Springer, 2015.
  • Roy and Kailath [1989]R.Roy and T.Kailath.ESPRIT-estimation of signal parameters via rotational invariance techniques.IEEE Transactions on Acoustics, Speech, and Signal Processing, 37(7):984–995, 1989.doi: 10.1109/29.32276.
  • Schmidt [1982]R.O. Schmidt.A signal subspace approach to multiple emitter location and spectral estimation.Stanford University, 1982.
  • Schumann etal. [2021]O.Schumann, M.Hahn, N.Scheiner, F.Weishaupt, J.F. Tilly, J.Dickmann, and C.Wöhler.RadarScenes: A real-world radar point cloud data set for automotive applications.arXiv preprint arXiv:2104.02493, 2021.
  • Sheeny etal. [2020]M.Sheeny, E.D. Pellegrin, M.Saptarshi, A.Ahrabian, S.Wang, and A.Wallace.RADIATE: A radar dataset for automotive perception.arXiv preprint arXiv:2010.09076, 2020.
  • Sun and Zhang [2021]S.Sun and Y.D. Zhang.4D automotive radar sensing for autonomous vehicles: A sparsity-oriented approach.IEEE Journal of Selected Topics in Signal Processing, 15(4):879–891, 2021.
  • Sun etal. [2020]S.Sun, A.P. Petropulu, and H.V. Poor.MIMO radar for advanced driver-assistance systems and autonomous driving: Advantages and challenges.IEEE Signal Processing Magazine, 37(4):98–117, 2020.
  • Texas Instruments [2020]Texas Instruments.Design Guide: TIDEP-01012 Imaging Radar Using Cascaded MmWave Sensor Reference Design, 2020.URL https://www.ti.com/lit/ug/tiduen5a/tiduen5a.pdf.Rev. A.
  • Waldschmidt etal. [2021]C.Waldschmidt, J.Hasch, and W.Menzel.Automotive radar — From first efforts to future systems.IEEE Journal of Microwaves, 1(1):135–148, 2021.
  • Wang etal. [2021]Y.Wang, Z.Jiang, X.Gao, J.-N. Hwang, G.Xing, and H.Liu.RODNet: Radar object detection using cross-modal supervision.In Proc. IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, Jan. 5-9, 2021.
  • Xue etal. [2011]M.Xue, L.Xu, and J.Li.IAA spectral estimation: Fast implementation using the Gohberg-sem*ncul factorization.IEEE Transactions on Signal Processing, 59(7):3251–3261, 2011.doi: 10.1109/TSP.2011.2131136.
  • Yardibi etal. [2010]T.Yardibi, J.Li, P.Stoica, M.Xue, and A.Baggeroer.Source localization and sensing: A nonparametric iterative adaptive approach based on weighted least squares.46(1):425–443, 2010.
  • Zheng etal. [2023]R.Zheng, S.Sun, H.Liu, and T.Wu.Deep-neural-network-enabled vehicle detection using high-resolution automotive radar imaging.IEEE Transactions on Aerospace and Electronic Systems, 59(5):4815–4830, 2023.
Redefining Automotive Radar Imaging: A Domain-Informed 1D Deep Learning Approach for High-Resolution and Efficient Performance (2024)
Top Articles
Latest Posts
Article information

Author: Fr. Dewey Fisher

Last Updated:

Views: 5469

Rating: 4.1 / 5 (42 voted)

Reviews: 89% of readers found this page helpful

Author information

Name: Fr. Dewey Fisher

Birthday: 1993-03-26

Address: 917 Hyun Views, Rogahnmouth, KY 91013-8827

Phone: +5938540192553

Job: Administration Developer

Hobby: Embroidery, Horseback riding, Juggling, Urban exploration, Skiing, Cycling, Handball

Introduction: My name is Fr. Dewey Fisher, I am a powerful, open, faithful, combative, spotless, faithful, fair person who loves writing and wants to share my knowledge and understanding with you.