Within the field of robotics, autonomous aerial and ground vehicles perform various tasks including navigation, mapping, or other sensing tasks. Currently, light detection and ranging (LiDAR) sensors, which have angular resolutions of 0.1 degrees and range resolutions of 2mm, are the gold standard for obtaining accurate and dense 3D point clouds of an environment. While they are lower resolution sensors, radio detection and ranging (radar) sensors provide accurate sensing while being up to 20x cheaper and consuming 3x less power. In this work, we introduce RadCloud, a novel real-time framework for directly obtaining higher resolution lidar-like 2D point clouds from low-resolution radar frames on resource-constrained platforms commonly used in unmanned aerial and ground vehicles. As described below, we enable real-time implementations by using a radar configuration with 1/4th the range resolution and 2.25x fewer parameters compared to previous radar implementations.
A common FMCW signal processing pipeline
Radio detection and ranging (a.k.a radar) sensors are commonly used in various applications due to their reliability in adverse lighting and weather conditions, long detection range, and ability to detect an object's relative velocity. While various techniques and waveforms can be used to perform radar sensing, frequency-modulated continuous wave (FMCW) radars are very common for low-cost sensing applications. The latest generation of FMCW radars, mm-Wave radar operates at frequencies ranging from 77-81GHz with bandwidths of up to 4 GHz; such high bandwidths and frequencies allow these sensors to have 20 x better range resolution (down to 4cm) and 3x better velocity resolution.
A common FMCW radar signal processing pipeline is featured above. From the figure, the 3 key steps of FMCW radar signal processing can be summarized as follows:
Chirps at Transmitter (Tx) and Receiver (Rx): For each radar frame, a series of chirps are transmitted. These chirps reflect off of objects in the environment, and the reflected chirps are received by the radar.
Dechirping and IF Signal Generation: The transmitted and received chirps are then fed through a mixer to determine a the intermediate frequency (a.k.a IF). This is the frequency difference between the transmitted and received chirp, and can easily be used to measure the range of different objects.
Range-Azimuth Response: Using an operation known as a 2D fast Fourrier transform (a.ka. 2D-FFT), we compute the Range-Azimuth response which shows the received signal power a various ranges and azimuth (horizontal) angles.
RadCloud Model Pipeline
We utilize a simplified U-net architecture in order to convert low resolution radar range-azimuth responses into higher resolution "lidar-like" 2D point clouds. The high level parameters of our model are described below:
Input: 40 x 64 x 48 normalized tensor corresponding to a normalized range-azimuth responses (in polar coordinates) from 40 chirps in a radar frame. For each chirp, we record 64 adc samples across 4 receivers. The azimuth FFT is zero padded to 64 elements, but we only keep the parts corresponding to +/- 50 degrees, resulting in 48 azimuth bin for each range
Output: 64 x 48 segmentation mask derived from the lidar ground truth data
Architecture: U-net
Loss Function: 0.9 weighted binary cross entropy (BCE) and 0.1 DICE loss
To demonstrate the real-world feasibility of RadCloud, we perform experiments using two real-time platforms. Here, the ground vehicle was used for model training and evaluation while the drone was used for real-time case studies. Below, we summarize the key details for each platform:
Unmanned Ground Vehicle (UGV):
Radar: TI-IWR1443 (radar sensor) and TI-DCA1000 (raw data streaming)
Lidar (for ground truth only): VLP-16 Puck
Computer: 2017 Intel NUC7i5BNH (no GPU)
Robot Platform: Kobuki ground vehicle
Unmanned Aerial Vehicle (AV):
Radar: TI-IWR1443 (radar sensor) and TI-DCA1000 (raw data streaming)
Computer: 2017 Intel NUC7i5BNH (no GPU)
Platform: DJI Matrice M1000