π PhD Candidate, Computer Science @ Old Dominion University
π Research Focus: Efficient ML Systems | LLM Inference Acceleration | Privacy-Preserving AI
β‘ Expertise: Quantization Β· Knowledge Distillation Β· Distributed Training Β· CUDA/GPU/TPU Optimization
π Security: Federated Learning Β· Secure Multi-Party Computation Β· Homomorphic Encryption
βοΈ Deployment: PyTorch Β· TensorFlow Β· AWS Β· GCP Β· Docker Β· KubernetesΒ· Triton
πΌ Industry Experience: AI/Cyber Research Intern @ SD Solutions LLC
2025. 9 One paper about robust personalized FL accepted by The Journal of Supercomputingπ₯³
2025. 8 One paper about model compression for private inference accepted by TMLR π₯³
2025. 6 One paper accepted by Computer Networksπ
2025. 5-8 I will intern at SD solutions. LLC. as AI/Cyber Research Internπ Thank you SD solutions.
2025. 1 Two papers accepted by Neurocomputing and Computer Communications
2024. 11 One paper about fast private Transformer inference is online.π
2024. 7 One paper about private inference on mobile computing accepted by ICDCS π§‘
π Designed parallel inference pipelines for privacy-preserving MLaaS, reducing latency on edge devices
β‘ Achieved ~60% faster inference with overlap scheduling and pipelining techniques
βοΈ collaborated with PyTorch + CUDA kernels, scalable to cloud deployment
βοΈ Demonstrated edge-to-cloud efficiency, bridging research and real-world deployment
Xu, X. et al. SPOT: Structure Patching and Overlap Tweaking for Effective Pipelining in Privacy-Preserving MLaaS with Tiny Clients. 2024 ICDCS.
Xu, J., Guan, C., & Xu, X. (2018). Energy-efficiency for smartphones using interaction link prediction in mobile cloud computing. CCF Conference on Computer Supported Cooperative Work and Social Computing.
βοΈClear out long-standing rumor in related research
ποΈ Developed quantization-friendly network linearization for secure inference
π Reduced communication overhead by ~50% without loss of accuracy
βοΈ Optimized LLM and vision model architectures for edge and cloud deployment
π Enabled model compression with minimal performance trade-offs
Xu, X. et al (2025). PrivShap: A Finer-granularity Network Linearization Method for Private Inference. TMLR
π Designed secure and communication-efficient transformer inference protocols
π Reduced communication cost by ~60% while maintaining model accuracy
π Improved throughput & latency for large-scale privacy-preserving inference
βοΈ Applied to LLMaaS and enterprise secure AI platforms, bridging research to deployment
Xu, X. et al (2024a). Comet: A communication-efficient and performant approximation for private transformer inference. ArXiv Preprint ArXiv:2405.17485.
β‘ Built resilient federated learning frameworks with dynamic trust adaptation and Byzantine robustness
π€ Developed personalized collaboration mechanisms for heterogeneous clients
π Introduced anti-forgetting strategies for incremental model updates
βοΈ Enabled robust and efficient distributed training across cloud and edge environments
Bai, Y., Wang, Y., Xu, X., Yang, Y., Batool, H., Iqbal, Z., & Xu, J. (2025). AsyncDefender: Dynamic trust adaptation and collaborative defense for Byzantine-robust asynchronous federated learning. Computer Networks, 111430.
Wang, Y., Xu, J., Yuan, Q., Yang, Y.,Β Xu, X., Batool, H. Fgcfl: a fine-grained clustering framework for federated learning with heterogeneity data. J Supercomput 81, 1328 (2025).Β
Xu, J., Zhao, Y., Li, X., Zhou, L., Zhu, K., Xu, X., Duan, Q., & Zhang, R. (2025). Teg-di: Dynamic incentive model for federated learning based on tripartite evolutionary game. Neurocomputing, 621, 129259.
Xu, J., Zhou, L., Zhao, Y., Li, X., Zhu, K., Xu, X., Duan, Q., & Zhang, R. (2025). A two-stage federated learning method for personalization via selective collaboration. Computer Communications, 232, 108053.
Zhu, K., Xu, J., Zhou, L., Li, X., Zhao, Y., Xu, X., & Li, S. (2025). Dmaf: data-model anti-forgetting for federated incremental learning. Cluster Computing, 28(1), 30.
Publications
Wang, Y., Xu, J., Yuan, Q., Yang, Y., Β Xu, X., Batool, H. Fgcfl: a fine-grained clustering framework for federated learning with heterogeneity data. J Supercomput 81, 1328 (2025).Β
Bai, Y., Wang, Y., Xu, X., Yang, Y., Batool, H., Iqbal, Z., & Xu, J. (2025). AsyncDefender: Dynamic trust adaptation and collaborative defense for Byzantine-robust asynchronous federated learning. Computer Networks, 111430.
Xu, X., Wang, Z., Ning, R., Xin, C., & Wu, H. (2025). PrivShap: A Finer-granularity Network Linearization Method for Private Inference. TMLR
Xu, J., Zhao, Y., Li, X., Zhou, L., Zhu, K., Xu, X., Duan, Q., & Zhang, R. (2025). Teg-di: Dynamic incentive model for federated learning based on tripartite evolutionary game. Neurocomputing, 621, 129259.
Xu, J., Zhou, L., Zhao, Y., Li, X., Zhu, K., Xu, X., Duan, Q., & Zhang, R. (2025). A two-stage federated learning method for personalization via selective collaboration. Computer Communications, 232, 108053.
Zhu, K., Xu, J., Zhou, L., Li, X., Zhao, Y., Xu, X., & Li, S. (2025). Dmaf: data-model anti-forgetting for federated incremental learning. Cluster Computing, 28(1), 30.
Xu, X., Zhang, Q., Ning, R., Xin, C., & Wu, H. (2024a). Comet: A communication-efficient and performant approximation for private transformer inference. ArXiv Preprint ArXiv:2405.17485.
Xu, X., Zhang, Q., Ning, R., Xin, C., & Wu, H. (2024b). SPOT: Structure Patching and Overlap Tweaking for Effective Pipelining in Privacy-Preserving MLaaS with Tiny Clients. 2024 IEEE 44th International Conference on Distributed Computing Systems (ICDCS), 1318β1329.
Garcia, K. R., Ammons, J., Xu, X., & Chen, J. (2023). Phishing in social media: Investigating training techniques on Instagram shop. Proceedings of the Human Factors and Ergonomics Society Annual Meeting, 67(1), 1850β1855.
Xu, J., Guan, C., & Xu, X. (2018). Energy-efficiency for smartphones using interaction link prediction in mobile cloud computing. CCF Conference on Computer Supported Cooperative Work and Social Computing, 517β526.
Reviewer: TNNLS, Information Science, CVPR 25, ICCV 25, ICLR 25