Shuteng Niu
Researcher
Department of Artificial Inrtelligence & Informatics
Mayo Clinic, Florida Campus
Office: Stabile, 750N
I am a researcher in the Department of Artificial Intelligence & Informatics at Mayo Clinic. I am also an adjunct faculty member at Bowling Green State University (BGSU), where I advise two Ph.D. students in Data Science. Before joining the Mayo Clinic, I was an assistant professor of computer science at BGSU and a postdoctoral research fellow at the School of Biomedical Informatics, UTHealth. I earned my Ph.D. in Computer Science from Embry-Riddle Aeronautical University.
My research areas include Transfer Learning (TL), Graph Representation Learning (GRL), and BioMedical Informatics (BMI). Currently, I am working on active and robust TL, continual GRL with dynamic regularization, and graph-driven drug repurposing.
Recent News
01/2025: Lijing Zhu and Dong Hyun Jeon passed their dissertation topic approval defenses! Congrats!
12/2024: I will join Mayo Clinic in February 2025, as a researcher in the Department of AI and Informatics.
12/2024: I will become an adjunct Assistant Professor at BGSU.
10/2024: Two papers from my Ph.D. students (Lijing Zhu and Dong Hyun Jeon) were accepted to IEEE BigData 2024.
10/2024: One collaborated paper was published at 2024 IEEE 36th International Conference on Tools with Artificial Intelligence (ICTAI).
10/2024: I gave a talk at INFORMS 2024.
07/2024: One collaborated paper was accepted to International Journal of Molecular Sciences.
02/2024: One collaborated paper was accepted to Journal of American Heart Association.
Featured Works
A decade survey of transfer learning (2010–2020) - IEEE Transactions on Artificial Intelligence 2021
Shuteng Niu, Yongxin Liu, Jian Wang, Houbing Song
Abstract: Transfer learning (TL) has been successfully applied to many real-world problems that traditional machine learning (ML) cannot handle, such as image processing, speech recognition, and natural language processing (NLP). Commonly, TL tends to address three main problems of traditional machine learning: (1) insufficient labeled data, (2) incompatible computation power, and (3) distribution mismatch. In general, TL can be organized into four categories: transductive learning, inductive learning, unsupervised learning, and negative learning. Furthermore, each category can be organized into four learning types: learning on instances, learning on features, learning on parameters, and learning on relations. This article presents a comprehensive survey on TL. In addition, this article presents the state of the art, current trends, applications, and open challenges.
Flexible Memory Rotation (FMR): Rotated Representation with Dynamic Regularization to Overcome Catastrophic Forgetting in Continual Knowledge Graph Learning - 2024 IEEE International Conference on Big Data 2024
Lijing Zhu, Dong Hyun Jeon, Wenbo Sun, Li Yang, Yixin Xie, Shuteng Niu
Abstract: Continual Knowledge Graph Learning (CKGL) is crucial in dynamic applications like recommendation systems and personalized services but faces the challenge of catastrophic forgetting. To address this, we propose Flexible Memory Rotation (FMR), a dual-level regularization technique inspired by human learning, which adapts constraints based on the quality of learned knowledge using the Fisher Information Matrix (FIM). FMR overcomes the limitations of traditional FIM assumptions by rotating the parameter space, leading to significant improvements over state-of-the-art CKGL models, as demonstrated on four benchmark datasets. Comprehensive experiments and ablation studies validate the effectiveness of FMR's components and its overall design.
KGIF: Optimizing Relation-Aware Recommendations with Knowledge Graph Information Fusion - 2024 IEEE International Conference on Big Data
Dong Hyun Jeon, Wenbo Sun, Houbing Herbert Song, Dongfang Liu, Alvaro Velasquez, Yixin Xie, Shuteng Niu
Abstract: Deep-learning-based recommender systems often struggle in real-world scenarios due to limited use of user-item relationship data and a lack of transparency. To address this, we propose the Knowledge Graph Attention Network with Information Fusion (KGIF), a framework that explicitly merges entity and relation embeddings using a tailored self-attention mechanism. By integrating dynamic projection vectors and attentive propagation, KGIF captures complex relationships in knowledge graphs, enhancing robustness in sparse settings and balancing user- and item-centric representations. Additionally, KGIF enables explainable recommendations through interpretable path visualization, offering a significant advance in both performance and transparency.
Spatial-Temporal Graph Data Mining for IoT-enabled Air Mobility Prediction - IEEE Internet of Things Journal 2021
Yushan Jiang, Shuteng Niu, Kai Zhang, Bowen Chen, Chengtao Xu, et al.
Abstract: We propose the first air traffic prediction paradigm based on spatial-temporal graph modeling. We provide a detailed study based on Airline On-Time Performance Data of the year 2016 (provided by Bureau of Transportation Statistics), and build an air transportation network (via different inductive biases) for 285 U.S. domestic airports with temporal features. Our framework based on spatial-temporal graph neural networks provides accurate joint predictions of the number, average delay, and average taxiing time of departure and arrival flights at each airport.
Main Collaborators
Dr. Wenbo Sun @ University of Michigan at Ann Arbor
Dr. Yixin Xie @ Kennesaw State University
Dr. Qing Tian @ University of Alabama at Birmingham
Kai Zhang @ Lehigh University
Yushang Jiang @ University of Connecticut
My old lab mates from SONG Lab