Publications


* Denotes equal contribution and co-first authors. 

_____ Denotes corresponding author.

Contactless Sensing with IOT for Smart Home and Healthcare

[25] Bing Zhu*, Zixin He*, Weiyi Xiong, Guanhua Ding, Jianan Liu, Tao Huang, Wei Chen, Wei Xiang, "ProbRadarM3F: mmWave Radar based Human Skeletal Pose Estimation with Probability Map Guided Multi-Format Feature Fusion", submitted to IEEE Transactions on Aerospace and Electronic Systems (T-AES), 2024.


Perception for ADAS and Autonomous Driving

[24] Guanhua Ding*, Jianan Liu*, Yuxuan Xia, Tao Huang, Bing Zhu, Jinping Sun, "When Extended Target Tracking Meets 3D Multi-Object Tracking with Point Clouds for Autonomous Driving", preparing to submit to IEEE Transactions on Intelligent Transportation Systems (T-ITS), 2024.

[23] Yanlong Yang*, Jianan Liu*, Tao Huang, "AutoDiff-Radar: Is Diffusion Model Really Good at Radar Point Cloud Enhancement for Perception of Automotive Environment?", preparing to submit to IEEE Robotics and Automation Letters (RA-L), 2024.

[22] Xi Zhou*, Jianan Liu*, Zhengrong Zhang, Tao Huang, Mostafa Rahimi Azghadi, "COPEIS: Cooperative Perception with Erroneous Information Sharing under Realistic Communication Environment", preparing to submit to IEEE Robotics and Automation Letters (RA-L), 2024.

[21] Runwei Guan*, Jianan Liu*, Ningwei Ouyang*, Haocheng Zhao, Liye Jia, Ka Lok Man, Ming Xu, Eng Gee Lim, Jeremy Smith, Yutao Yue, "USV-Centric Low-Power Multi-Task Visual Grounding based on Prompt-Guided Camera and 4D mmWave Radar", preparing to submit to The 50th IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2025.


[20] Runwei Guan*, Ruixiao Zhang*, Ningwei Ouyang*, Jianan Liu*, Ka Lok Man, Ming Xu, Jeremy Smith, Eng Gee Lim, Yutao Yue, Hui Xiong, "Talk2Radar: Bridging Natural Language with 4D mmWave Radar for 3D Referring Expression Comprehension", submitted to The 42th IEEE International Conference on Robotics and Automation (ICRA), 2025. Code

Task: Embodied Perception and Multi-Modal Large Language Model

Embodied perception is essential for intelligent vehicles and robots, enabling more natural interaction and task execution. However, these advancements currently embrace vision level, rarely focusing on using 3D modeling sensors, which limits the full understanding of surrounding objects with multigranular characteristics. Recently, as a promising automotive sensor with affordable cost, 4D Millimeter-Wave radar provides denser point clouds than conventional radar and perceives both semantic and physical characteristics of objects, thus enhancing the reliability of perception system. To foster the development of natural language-driven context understanding in radar scenes for 3D grounding, we construct the first dataset, Talk2Radar, which bridges these two modalities for 3D Referring Expression

Comprehension. Talk2Radar contains 8,682 referring prompt samples with 20,558 referred objects. Moreover, we propose a novel model, T-RadarNet for 3D REC upon point clouds, achieving state-of-the-art performances on Talk2Radar dataset compared with counterparts, where Deformable-FPN and Gated Graph Fusion are meticulously designed for efficient point cloud feature modeling and cross-modal fusion between radar and text features, respectively. Further, comprehensive experiments are conducted to give a deep insight into radar-based 3D REC. We release our project at https://github.com/GuanRunwei/Talk2Radar.

[19] Zhenrong Zhang*, Jianan Liu*, Xi Zhou, Tao Huang, Qing-Long Han, Hongbin Liu, "On the Federated Learning Framework for Cooperative Perception", submitted to IEEE Robotics and Automation Letters (RA-L), 2024.

Task: Federated Learning Algorithm Design for Cooperative Perception

Cooperative perception is essential to enhance the efficiency and safety of future transportation systems, requiring extensive data sharing among vehicles on the road, which raises significant privacy concerns. Federated learning offers a promising solution by enabling data privacy-preserving collaborative enhancements in perception, decision-making, and planning among connected and autonomous vehicles (CAVs). However, federated learning is impeded by significant challenges arising from data heterogeneity across diverse clients, potentially diminishing model accuracy and prolonging convergence periods. This study introduces a specialized federated learning framework for C-ITS, termed the federated dynamic weighted aggregation (Fed-DWA) algorithm, facilitated by dynamic adjusting loss (DALoss) function. This framework employs dynamic client weighting to direct model convergence and integrates a novel loss function

that utilizes Kullback-Leibler divergence (KLD) to counteract the detrimental effects of non-independently and identically distributed (Non-IID) and unbalanced data. Utilizing the BEV transformer as the primary model, our rigorous testing on the OpenV2V dataset, augmented with FedBEVT data, demonstrates significant improvements in the average intersection over union (IoU). These results highlight the substantial potential of our federated learning framework to address data heterogeneity challenges in C-ITS, thereby enhancing the accuracy of environmental perception models and facilitating more robust and efficient collaborative learning solutions in the transportation sector.

[18] Guanhua Ding, Jianan Liu, Yuxuan Xia, Tao Huang, Bing Zhu, Jinping Sun, "LiDAR Point Cloud-based Multiple Vehicle Tracking with Probabilistic Measurement-Region Association", The 27th International Conference on Information Fusion (FUSION), 2024.

Task: LiDAR Point Cloud-based Multiple Extended Target Tracking

Multiple target tracking (MTT) is a crucial element in autonomous driving and intelligent transportation systems. Recently, extended target tracking (ETT) has aroused attention in the context of point cloud-based MTT, and several ETT methods based on data-region association (DRA) have been proposed for tracking vehicle targets. The DRA methods can accurately describe the uneven distribution of point clouds by dividing the target extent into separate regions. However, the constrained estimation required by DRA methods increases complexity and approximation errors, while the accuracy and stability for calculating association probabilities are affected by the randomly distributed scattering center. In this paper, we propose a novel ETT algorithm based on probabilistic measurement-region association (PMRA) and Poisson multi-Bernoulli mixture (PMBM) filter for tracking multiple vehicle targets with LiDAR point clouds. The PMRAPMBM algorithm addresses the limitations of DRA model by eliminating the estimation constraint and calculating association probabilities with continuous integrals. Simulation results show the superior estimation accuracy of PMRA-PMBM on target location and extent comparing with the gamma Gaussian inverse Wishart and DRA implementations of the PMBM filter.

[17] Tao Huang*, Jianan Liu*, Xi Zhou*, Dinh C. Nguyen, Mostafa Rahimi Azghadi, Yuxuan Xia, Qing-Long Han, Sumei Sun, "V2X Cooperative Perception for Autonomous Driving: Recent Advances and Challenges", submitted to Proceedings of the IEEE (P-IEEE), 2024.

Task: Survey of Cooperative Perception with V2X Communication for ADAS and AD

This paper provides a comprehensive overview of the evolution of cooperative perception (CP) technologies, spanning from early explorations to recent developments until Aug. 2023, including advancements in V2X communication technologies. Additionally, we propose a contemporary generic framework to illustrate the V2X cooperation workflow, aiding in the structured understanding of CP system components. Furthermore, this paper categorizes prevailing V2X CP methodologies based on the critical issues they address. We conduct an extensive literature review within this taxonomy, evaluating existing datasets and simulators. Finally, we discuss open challenges and future directions in CP for autonomous driving, considering both perception and V2X communication advancements.

[16] Jianan Liu*, Guanhua Ding*, Yuxuan Xia, Jinping Sun, Tao Huang, Lihua Xie, Bing Zhu, "Which Framework is Suitable for Online 3D Multi-Object Tracking for Autonomous Driving with Automotive 4D Imaging Radar?", The 35th IEEE Intelligent Vehicles Symposium (IV), 2024. Oral pesentations, Top 5%. Code

Task: 3D Multi-Object Tracking (MOT) with 4D Imaging Radar

This paper provides the first systematical investigation of the EOT framework for online 3D MOT in real-world ADAS and AD scenarios. Specifically, the widely accepted TBD-POT framework, the recently investigated JDT-EOT framework, and our proposed TBD-EOT framework are compared via extensive evaluations on two open source 4D imaging radar datasets: View-of-Delft and TJ4DRadSet. These provide the first benchmark and important insights for the future development of 4D imaging radar-based online 3D MOT

[15] Zhenrong Zhang*, Jianan Liu*, Yuxuan Xia, Tao Huang, Qing-Long Han, Hongbin Liu, "LEGO: Learning and Graph-Optimized Modular Tracker for Online Multi-Object Tracking with Point Clouds", submitted to IEEE Transactions on Vehicular Technology (T-VT), 2024.

Task: 3D Multi-Object Tracking (MOT) with LiDAR Point Cloud

This paper proposes a learning and graph-optimized (LEGO) modular tracker to improve data association performance in the existing literature. The proposed LEGO tracker integrates graph optimization and self-attention mechanisms, which efficiently formulate the association score map, facilitating the accurate and efficient matching of objects across time frames. LEGO ranked 1st at the time of submitting results to KITTI object tracking evaluation ranking board and remains 2nd at the time of submitting this paper, among all online trackers in the KITTI MOT benchmark for cars

[14] Yanlong Yang*, Jianan Liu*, Tao Huang, Qing-Long Han, Gang Ma, Bing Zhu, "RaLiBEV: Radar and LiDAR BEV Fusion Learning for Anchor Box Free Object Detection Systems", submitted to IEEE Transactions on Intelligent Transportation Systems (T-ITS), 2024.

Task: 3D Object Detection with Radar and LiDAR Fusion

In this paper, we propose a bird’seye view fusion learning-based anchor box-free object detection system, which fuses the feature derived from the radar rangeazimuth heatmap and the LiDAR point cloud to estimate the possible objects. Different label assignment strategies have been designed to facilitate the consistency between the classification of foreground or background anchor points and the corresponding bounding box regressions. In addition, the performance of the proposed object detector is further enhanced by employing a novel interactive transformer module. The superior performance of the methods proposed in this paper has been demonstrated using the recently published Oxford Radar RobotCar dataset. Our system is compared with the apporaches proposed for the same task with same dataset in CVPR 2020, 2021, 2022, 2023 and ICASSP 2023, significantly outperforms the best one among them by 13.1% and 19.0% at IoU of 0.8 under ’Clear+Foggy’ training conditions for ’Clear’ and ’Foggy’ testing, respectively, thus serves as the latest benchmark for radar and LiDAR fusion-based BEV object detection.

[13] Weiyi Xiong*, Jianan Liu*, Tao Huang, Qing-Long Han, Yuxuan Xia, Bing Zhu, "LXL: LiDAR Excluded Lean 3D Object Detection with 4D Imaging Radar and Camera Fusion", IEEE Transactions on Intelligent Vehicles (T-IV), vol. 9, no. 1, pp. 79 - 92, 2024. ESI Highly Cited Paper.

Task: 3D Object Detection with 4D Imaging Radar and Camera Fusion

In this paper, we investigate the “sampling” view transformation strategy on the camera and 4D imaging radar fusion-based 3D object detection. In the proposed LiDAR Excluded Lean (LXL) model, predicted image depth distribution maps and radar 3D occupancy grids are generated from image perspective view (PV) features and radar bird’s eye view (BEV) features, respectively. They are sent to the core of LXL, called “radar occupancy-assisted depth-based sampling”, to aid image view transformation. Introducing image depths and radar information enhances the “sampling” strategy and leads to more accurate view transformation. Experiments on VoD and TJ4DRadSet datasets show that the proposed method outperforms the stateof-the-art 3D object detection methods (e.g., FUTR3D(CVPR 2023), BEVFusion(ICRA 2023), RCFusion(IEEE T-IM), RCN(ICCV 2023) with 4D imaging radar and camera fusion settings) by a significant margin without bells and whistles. Ablation studies demonstrate that our method performs the best among different enhancement settings. This is an early attempt in this field, and serves as the latest benchmark for subsequent studies

[12] Jianan Liu*, Qiuchi Zhao*, Weiyi Xiong, Tao Huang, Qing-Long Han, Bing Zhu, "SMURF: Spatial Multi-Representation Fusion for 3D Object Detection with 4D Imaging Radar”, IEEE Transactions on Intelligent Vehicles (T-IV), vol. 9, no. 1, pp. 799 - 812, 2024. ESI Highly Cited Paper.

Task: 3D Object Detection with 4D Imaging Radar

This paper introduces spatial multi-representation fusion (SMURF), a novel approach to 3D object detection using a single 4D imaging radar. SMURF leverages multiple representations of radar detection points, including pillarization and density features of a multidimensional Gaussian mixture distribution through kernel density estimation (KDE). Experimental evaluations on View-of-Delft (VoD) and TJ4DRadSet datasets demonstrate the effectiveness and generalization ability of SMURF, outperforming recently proposed 4D imaging radarbased single-representation models. Moreover, while using 4D imaging radar only, SMURF still achieves comparable performance to the state-of-the-art 4D imaging radar and camera fusion-based method, with an increase of 1.22% in the mean average precision on bird’s-eye view of TJ4DRadSet dataset and 1.32% in the 3D mean average precision on the entire annotated area of VoD dataset. This research highlights the benefits of 4D mmWave radar and is a latest benchmark for subsequent works regarding 3D object detection with 4D imaging radar

[11] Jianan Liu*, Liping Bai*, Yuxuan Xia, Tao Huang, Bing Zhu, Qing-Long Han, "GNN-PMB: A Simple But Effective Online 3D Multi-Object Tracker without Bells and Whistles", IEEE Transactions on Intelligent Vehicles (T-IV), vol. 8, no. 2, pp. 1176-1189, 2023. Code

Task: 3D Multi-Object Tracking (MOT) with LiDAR

In this paper, it is demonstrated that the latest RFS-based Bayesian tracking framework could be superior to typical random vector-based Bayesian tracking framework via a systematic comparative study of both traditional random vector-based Bayesian filters with rule-based heuristic track maintenance and RFS-based Bayesian filters on the nuScenes validation dataset. An RFS-based tracker, namely Poisson multiBernoulli filter using the global nearest neighbor (GNN-PMB), is proposed to LiDAR-based MOT tasks. This GNN-PMB tracker is simple to use, and it achieves competitive results on the nuScenes dataset. Specifically, the proposed GNN-PMB tracker outperforms most state-of-the-art LiDAR-only trackers and LiDAR and camera fusion-based trackers, ranking the 3rd among all LiDAR-only trackers on nuScenes 3D tracking challenge leader board at the time of submission

[10] Weiyi Xiong*, Jianan Liu*, Yuxuan Xia, Tao Huang, Bing Zhu, Wei Xiang, "Contrastive Learning for Automotive mmWave Radar Detection Points based Instance Segmentation", The 25th IEEE International Conference on Intelligent Transportation Systems (ITSC), pp. 1255-1261, 2022.

Task: Instance Segmentation with Automotive Radar with Contrastive Learning

To solve the issue that high-quality annotations of radar detection points are challenging to achieve due to their ambiguity and sparsity, we propose a contrastive learning approach for implementing radar detection points-based instance segmentation. Experiments show that when the ground-truth information is only available for a small proportion of the training data, our method still achieves a comparable performance to the approach trained in a supervised manner with 100% ground-truth information.

[9] Jianan Liu*, Weiyi Xiong*, Liping Bai, Yuxuan Xia, Tao Huang, Wanli Ouyang, Bing Zhu, "Deep Instance Segmentation with Automotive Radar Detection Points", IEEE Transactions on Intelligent Vehicles (T-IV), vol. 8, no. 1, pp. 84-94, 2023. ESI Highly Cited Paper.

Task: Instance Segmentation with Automotive Radar

In this paper, we propose an efficient method based on clustering of estimated semantic information to achieve instance segmentation for the sparse radar detection points. In addition, we show that the performance of the proposed approach can be further enhanced by incorporating the visual multi-layer perceptron. The effectiveness of the proposed method is verified by experimental results on the popular RadarScenes dataset, achieving 89.53% mean coverage and 86.97% mean average precision with the IoU threshold of 0.5, which is superior to other approaches in the literature. More significantly, the consumed memory is around 1MB, and the inference time is less than 40ms, indicating that our proposed algorithm is storage and time efficient. These two criteria ensure the practicality of the proposed method in real-world systems

Deep Learning and Signal Processing for Medical Imaging in Clinical Setting

[8] Hao Li*, Yusheng Zhou*, Jianan Liu, Xiling Liu, Tao Huang, "KINRUM: K-Space Implicit Neural Representation for Unsupervised 3D MRI Reconstruction in Practical Clinics", preparing to submit to IEEE Journal of Biomedical and Health Informatics (J-BHI), 2024.

[7] Hao Li*, Quanwei Liu*, Jianan Liu, Xiling Liu, Yanni Dong, Tao Huang, Zhihan Lv, "Unpaired 3D MRI Super Resolution with Contrastive Learning", The 21st IEEE International Symposium on Biomedical Imaging (ISBI), 2024.

Task: Unsupervised Learning MRI Super Resolution with Limited HR MRI Images

Magnetic resonance imaging (MRI) is crucial for enhancing diagnostic accuracy in clinical settings. However, the inherent long scan time of MRI restricts its widespread applicability. Deep learning-based image super-resolution (SR) methods exhibit promise in improving MRI resolution without additional cost. Due to lacking of aligned high-resolution (HR) and low-resolution (LR) MRI image pairs, unsupervised approaches are widely adopted for SR reconstruction with unpaired MRI images. However, these methods still require a substantial number of HR MRI images for training, which can be difficult to acquire. To this end, we propose an unpaired MRI SR approach that employs contrastive learning to enhance SR performance with limited HR training data. Empirical results presented in this study underscore significant enhancements in the peak signal-to-noise ratio and structural similarity index, even when a paucity of HR images is available. These findings accentuate the potential of our approach in addressing the challenge of limited HR training data, thereby contributing to the advancement of MRI in clinical applications. 


[6] Hao Li*, Yusheng Zhou*, Jianan Liu, Xiling Liu, Tao Huang, Zhihan Lv, Weidong Cai, "Implicit Neural Representation for MRI Parallel Imaging Reconstruction", submitted to IEEE Journal of Biomedical and Health Informatics (J-BHI), 2024.

Task: Supervised Learning MRI Parallel Reconstruction with Implicit Neural Representation

In this paper, we propose a novel MRI reconstruction method based on INR, which represents the fully-sampled images as the function of pixel coordinates and prior feature vectors of undersampled images for overcoming the generalization problem of INR. Specifically, we introduce a scale-embedded encoder to produce scale-independent pixel-specific features from MR images with different undersampled scales and then concatenate with coordinates vectors to recover fully-sampled MR images via an MLP, thus achieving arbitrary scale reconstruction. The performance of the proposed method was assessed by experimenting on publicly available MRI datasets and compared with other reconstruction methods. Our quantitative evaluation demonstrates the superiority of the proposed method over alternative reconstruction methods. 

[5] Yusheng Zhou*, Hao Li*, Jianan Liu, Zhengmin Kong, Tao Huang, Euijoon Ahn, Zhihan Lv, Jinman Kim, David Dagan Feng, "UNAEN: Unsupervised Abnomality Extraction Network for MRI Motion Artifact Reduction", submitted to IEEE Journal of Biomedical and Health Informatics (J-BHI), 2024.

Task: Unsupervised Learning MRI Motion Artifact Reduction

One disadvantage of supervised learning methods for MRI motion artifact reduction is their dependency on acquiring paired sets of motion artifact-corrupted (MA-corrupted) and motion artifact-free (MA-free) MR images for training purposes. Obtaining such image pairs is difficult in real clinics and therefore limits the application of supervised training. In this paper, we propose a novel UNsupervised Abnormality Extraction Network (UNAEN) which is capable of working with unpaired MA-corrupted and MA-free images. It converts the MA-corrupted images to MA-reduced images by extracting abnormalities from the MA-corrupted images using a proposed artifact extractor, which intercepts the residual artifact maps from the MA-corrupted MR images explicitly, and a reconstructor to restore the original input from the MA-reduced images. The performance of UNAEN was assessed by experimenting on various publicly available MRI datasets and comparing them with state-ofthe-art methods. The quantitative evaluation demonstrates the superiority of UNAEN over alternative MAR methods and visually exhibits fewer residual artifacts. Our results substantiate the potential of UNAEN as a promising solution applicable in real-world clinical environments, with the capability to enhance diagnostic accuracy and facilitate image-guided therapies

[4] Jianan Liu*, Hao Li*, Tao Huang, Kang Han, Euijoon Ahn, Adeel Razi, Wei Xiang, Jinman Kim, David Dagan Feng, "Unsupervised Representation Learning for 3D MRI Super Resolution with Degradation Adaptation", IEEE Transactions on Artificial Intelligence (T-AI), 2024.

Task: Unsupervised Learning MRI Super Resolution

Training neural networks for MRI super resolution application in supervised learning strategy requires aligned authentic HR and LR image pairs, which are challenging to obtain due to patient movements during and between image acquisitions. While rigid movements of hard tissues can be corrected with image registration, aligning deformed soft tissues is complex, making it impractical to train neural networks with authentic HR and LR image pairs. Previous studies have focused on SRR using authentic HR images and down-sampled synthetic LR images. However, the difference in degradation representations between synthetic and authentic LR images suppresses the quality of SR images reconstructed from authentic LR images. To address this issue, we propose a novel Unsupervised Degradation Adaptation Network (UDEAN). Our network consists of a degradation learning network and an SRR network. The degradation learning network downsamples the HR images using the degradation representation learned from the misaligned or unpaired LR images. The SRR network then learns the mapping from the down-sampled HR images to the original ones. Experimental results show that our method outperforms state-of-the-art networks and is a promising solution to the challenges in clinical settings

[3] Hao Li*, Jianan Liu*, Tim Hilgenfeld, Sabine Heiland, Martin Bendszus, "3D High-Quality Magnetic Resonance Image Restoration in Clinics Using Deep Learning", submitted to Journal of Magnetic Resonance Imaging (J-MRI), 2023.

Task: Supervised Learning MRI Super Resolution with Uncertainty Estimating

We adopted a unified framework of 2D deep learning neural network for both 3D MRI super-resolution and motion artifact reduction, demonstrating such a framework can achieve better performance in 3D MRI restoration tasks compared to other stateof-the-art methods and remain the GPU consumption and inference time significantly low, thus easier to deploy. We also analyzed several down-sampling strategies based on the acceleration factor, including multiple combinations of in-plane and through-plane down-sampling, and developed a controllable and quantifiable motion artifact generation method. At last, the pixel-wise uncertainty was calculated and used to estimate the accuracy of the generated image, providing additional information for a reliable diagnosis

[2] Sowmya Annadatha, Qirui Hua, Marie Fridberg, Tobias Lindstrøm Jensen, Jianan Liu, Søren Kold, Ole Rahbek, Ming Shen, "Preparing Infection Detection Technology for Hospital at Home after Lower Limb External Fixation", Digital Health, volume 8, pp. 1–20, 2022.

Wireless Communication

[1] Xiang Gao, Ove Edfors, Jianan Liu, Fredrik Tufvesson, "Antenna Selection in Measured Massive MIMO Channels using Convex Optimization", The 56th IEEE Global Communications Conference(Globecom) Workshop on Emerging Technologies for LTE-Advanced and Beyond-4G, pp. 129-134, 2013.