The Next Talk is on Nov. 10 by
Armita KazemiNajafabadi, Northeastern University
Zoom registration:https://cityu.zoom.us/meeting/register/KfTefDbQTNa9C9qDgNNpWw
If you have any difficulty registering, please contact taoli96@ieee.org.
Schedule At a Glance
Nov. 10, 2025, 2:00–3:00 pm UTC (6–7 am US West, 9–10 am US East, 2–3 pm UK, 3–4 pm CET, 7:30–8:30 pm India, 10–11 pm China)
Abstract: Multi-agent reinforcement learning (MARL) has emerged as a promising approach for adaptive and scalable cyber defense, enabling distributed agents to learn coordinated defense policies in complex and uncertain network environments. Its ability to adaptively optimize decisions across multiple agents makes MARL appealing for defending against evolving cyber threats. In a broader area of artificial intelligence, existing adversarial examples have shown that even sophisticated models can be misled, causing classifiers or learned policies to make incorrect decisions. This raises important questions for cyber defense: could similar adversarial strategies undermine MARL-based defenders, and how resilient are they when facing intelligent and coordinated deception?
In our work, we investigate the design of AI-powered adversaries that challenge MARL defense policies in distributed network environments. By modeling defenders’ interactions as a decentralized decision-making process under uncertainty, we develop new classes of adversarial strategies that intelligently manipulate feedback, disrupt information flow, and interfere with coordination—operating under both resource and stealth constraints to systematically degrade groups of AI decision makers. Our findings reveal critical gaps in MARL defense mechanisms and motivate next-generation security frameworks that explicitly account for deception-aware adversaries, advancing the robustness of AI-driven cyber defense in dynamic and adversarial environments.
Bio: Armita KazemiNajafabadi is a Ph.D. candidate in Electrical and Computer Engineering at Northeastern University, advised by Prof. Mahdi Imani. She received her B.Sc. in Computer Science with a minor in Mathematics from Sharif University of Technology, Tehran, Iran, in 2022, and joined Northeastern with a scholarship as an outstanding Ph.D. student. Her research lies at the intersection of reinforcement learning, game theory, and cybersecurity, with a particular focus on securing cyber-physical systems against evolving adversarial threats. Her recent work develops frameworks for adversarial cyber-attacks on deep learning-based defenders, Bayesian attack graph modeling, and defense strategies under uncertainty. Armita’s contributions have been published in leading journals and presented at major conferences such as the American Control Conference (ACC), IEEE CASE, AIAA SciTech, and the SIGCOMM N2Women Workshop. Her research has also been recognized with the People’s Choice Award from the Northeastern University College of Engineering PhD Council for her impactful work on network security. Her broader research interests include reinforcement learning, graphical models, and intelligent adversarial modeling, with the long-term goal of shaping next-generation AI-driven defense mechanisms for critical cyber-physical systems.
Nov. 21, 2025, 2:00–3:00 pm UTC (6–7 am US West, 9–10 am US East, 2–3 pm UK, 3–4 pm CET, 7:30–8:30 pm India, 10–11 pm China)
Abstract: This talk explores how principles of cyber-physical system (CPS) security and resilience can be effectively translated into the design of secure, autonomous, and self-healing microservices within the FinTech sector. Drawing from real-world industry experience, I will demonstrate how agentic AI, performance-aware microservices, and zero-trust principles can be used to safeguard high-volume financial transactions from disruptions, anomalies, and threats. The session will cover practical architectures for self-tuning systems, discuss failure domains in distributed transaction systems, and present secure-by-design strategies for cloud-native applications. By aligning industry-grade systems with CPS-like resilience models, this talk aims to foster discussion between theoretical control systems security and its real-world application in modern digital finance infrastructure.
Bio: Sibasis Padhi is a Staff Software Engineer at Walmart Global Tech, specializing in AI-powered microservices and cloud performance optimization for financial applications. With over 18 years of experience spanning telecom, cloud, and FinTech systems, he has led large-scale engineering initiatives that modernize secure transaction processing and financial planning infrastructure. His work bridges the gap between real-time reliability and scalable architecture by embedding resilience and intelligent tuning in cloud-native systems. A senior member of IEEE and Fellow of ICETAIA, Sibasis regularly speaks at international conferences and mentors engineers on building secure, high-performance systems. His ongoing work involves applying agentic AI and autonomous decision frameworks to improve the trustworthiness and uptime of critical financial workflows.
Nov. 25, 2025, 2:00–3:00 pm UTC (6–7 am US West, 9–10 am US East, 2–3 pm UK, 3–4 pm CET, 7:30–8:30 pm India, 10–11 pm China)
Abstract: Modern societies increasingly rely on AI-driven cyber-physical-human systems (CPHSs), such as intelligent transportation, industrial automation, and other critical infrastructure. While these systems promise efficiency and intelligence, they also introduce new vulnerabilities where security, privacy, and resilience are tightly coupled with human trust. A central question arises: how can we design socio-technical systems that remain trustworthy and resilient even in the presence of adversarial manipulation and the cognitive biases of human decision-makers? In this talk, we will present a research agenda that develops principled and computationally tractable frameworks for understanding trust in CPHSs. We will walk through four key perspectives: assessing trust via meta-game analysis of human–CPS interactions, building trust in AI through crowd auditing and accountability mechanisms, exploiting trust in adversaries through defensive deception, and maintaining user trust under misinformation with information design strategies. These frameworks will be illustrated through case studies in critical CPHS domains. We will conclude by outlining future directions toward resilient, cognitive-aware CPHSs.
Bio: Ya-Ting Yang is a Ph.D. candidate in Electrical and Computer Engineering at New York University, affiliated with the NYU Center for Cybersecurity. She received her M.S. in Communication Engineering from National Taiwan University and her B.S. in Electrical Engineering from National Tsing Hua University. Her research focuses on game theory and optimization with applications to the security and resilience of AI-driven cyber-physical-human systems. She is an RSAC Security Scholar, and her contributions have been published in leading journals and conferences such as IEEE TIFS, TNSE, TNSM, TITS, IOTJ, CDC, and Globecom.
Dec. 17, 2025, 2:00–3:00 pm UTC (6–7 am US West, 9–10 am US East, 2–3 pm UK, 3–4 pm CET, 7:30–8:30 pm India, 10–11 pm China)
Abstract: Bayesian data fusion offers a principled route to distributed learning under privacy and uncertainty. This talk develops a unifying framework that clarifies how local beliefs should be combined when priors are shared. We analyze the Conditionally Independent Likelihood (CIL) and Conditionally Independent Posterior (CIP) rules, identify the prior double-counting pitfall in naïve posterior multiplication, and derive corrections that preserve coherence while characterizing accuracy as a function of client count and prior informativeness, beyond Gaussian models. Building on this foundation, we introduce federated posterior sharing for multi-agent systems, in which agents exchange posteriors rather than data to construct a global belief and act. The method supports single-shot or periodic synchronization, avoids prior reuse, and improves reward and sample efficiency under uncertainty and heterogeneity. Finally, we present a Bayesian formulation of clustered federated learning that treats client–cluster assignment as latent data association, yielding practical approximations that handle non-IID feature and label skew and outperform standard clustered FL. Together, these results provide a coherent recipe—fuse beliefs, correct for shared priors, and quantify uncertainty—for privacy-preserving learning and decision making at scale.
Bio: Peng Wu is a Postdoctoral Researcher in the Department of Electrical and Computer Engineering at Northeastern University under the advice of Professor Mahdi Imani. His research focuses on security, privacy, and trustworthiness in distributed intelligence, with applications to mixed-reality (cognitive) attacks, multi-agent and robotic collaboration, and privacy-preserving machine learning. He develops methods at the intersection of machine learning, reinforcement learning, and federated learning to enable robust and reliable intelligent systems. He received his Ph.D. in Electrical Engineering from Northeastern University under the supervision of Professor Pau Closas, where his doctoral work advanced Bayesian and deep learning approaches for federated learning, applied to indoor positioning, Global Navigation Satellite Systems (GNSS), and image processing. He also holds a Master’s degree in Electrical Engineering at Northeastern University. His work has led to publications in leading journals and conferences, including IEEE Transactions on Signal Processing, IEEE Transactions on Aerospace and Electronic Systems, and the IEEE/ION Position, Location and Navigation Symposium (PLANS), where he received a Best Paper Award in 2023. Beyond academia, Peng co-founded Cactivate, an AI-driven platform that optimizes and automates online advertising for agencies and e-commerce businesses, serving as its Chief Scientist.