The research areas of current interest include (but are not limited to): Edge AI, Online learning, Age of Information (AoI), Delay analysis in networks.
Efficient machine learning Inference in Edge AI
Using Deep Learning (DL) models to do inference on data generated by end devices such as smartphones, IoT sensors and drones results in significant optimizations and efficiency gains across industrial automation, smart cities, remote healthcare, smart agriculture, and other sectors. In contrast to doing the inference by collecting data centrally at a datacenter, doing it on devices close to the data generation reduces response time and network bandwidth usage, enhances privacy, and improves scalability. However, the tiny DL models that can be embedded on the devices are unreliable, as they have lower accuracy compared to the large DL models, and they cannot generalize well. The overarching goal of my research is to make on-device DL inference reliable, robust, and energy efficient. Please take a look at the following works.
G Al-Atat, P Datta, S Moharir, JP Champati, "Regret Bounds for Online Learning for Hierarchical Inference," ACM MobiHoc, 2024.
VN Moothedath, JP Champati, J. Gross, "Getting the Best Out of Both Worlds: Algorithms for Hierarchical Inference at the Edge," IEEE Transactions on Machine Learning in Communications and Networking (TMLCN), 2024.
AP Behera, P Daubaris, I Bravo, J Gallego, R Morabito, J Widmer, JP Champati, "Exploring the Boundaries of On-Device Inference: When Tiny Falls Short, Go Hierarchical," arXiv:2407.11061, 2024.
Adarsh Prasad Behera, Roberto Morabito, Joerg Widmer, and Jaya Prakash Champati, Improved Decision Module Selection for Hierarchical Inference in Resource-Constrained Edge Devices, MobiCom 2023.
Ghina Al-Atat, Andrea Fresa, Adarsh Prasad Behera, Vishnu Narayanan Moothedath, James Gross, and Jaya Prakash Champati, The Case for Hierarchical Deep Learning Inference at the Network Edge, NetAI workshop, MobiSys 2023.
Andrea Fresa, Jaya Prakash Champati, Offloading Algorithms for Maximizing Inference Accuracy on Edge Device Under a Time Constraint, IEEE TPDS, 2023.
AoI Analysis and Optimization
AoI is a freshness metric that measures the time elapsed since the generation time of the freshest packet available at the Receiver. In contrast to system delay, AoI increases linearly between the packet receptions through which it accounts for the frequency of sampling the Information Source. We analyze AoI for fundamental queueing systems and also study optimal sampling and transmission strategies for minimizing AoI in these systems.
M Fidler, JP Champati, J Widmer, M Noroozi, "Statistical age-of-information bounds for parallel systems: When do independent channels make a difference?" IEEE Journal on Selected Areas in Information Theory, 2023
Jaya Prakash Champati, Hussein Al-Zubaidy, James Gross. Statistical Guarantee Optimization for AoI in Single-Hop and Two-Hop FCFS Systems with Periodic Arrivals. IEEE Transactions on Communications. 69 - 1, pp. 365 - 381. 2021.
Jaya Prakash Champati, Ramana R. Avula, Tobias J. Oechtering; James Gross. Minimum Achievable Peak Age of Information Under Service Preemptions and Request Delay. IEEE Journal on Selected Areas in Communications. 39 - 5, pp. 1365 - 1379. 2021.
Jaya Prakash Champati and Hussein Al-Zubaidy and James Gross, “On the distribution of AoI for the GI/GI/1/1 and GI/GI/1/2* systems: Exact expressions and bounds”, in Proc. IEEE INFOCOM, May 2019.
Transient Delay Analysis and Optimization
Most of the research dealing with general network performance analysis using queuing theory consider systems in steady-state. For example, for simple M/M/1 or more general Markovian queuing systems, the steady-state is governed by the (conceptually simple) flow balance equations. In contrast, transient analysis of these systems results in intractable differential equations. Using Stochastic Network Calculus, we derived the end-to-end delay violation probability for a sequence of time-critical packets, given the transient network state (queue backlogs) when the time-critical packets enter the network. Leveraging this analysis we compute good resource allocation strategies for wireless protocols such as WirelessHART to support the QoS requirements of time-critical industrial applications.
S Zoppi, JP Champati, J Gross, W Kellerer, "Scheduling of wireless edge networks for feedback-based interactive applications," IEEE Transactions on Communications, 2022.
Jaya Prakash Champati, Hussein Al-Zubaidy, James Gross. "Transient Analysis for Multi-hop Wireless Networks Under Static Routing," IEEE/ACM Transactions on Networking, 2020.
Edge Computing Offloading Algorithms
Edge computing or fog computing, where computational resources are placed close to (e.g. one hop away) entities that offload computational tasks or data for processing, is a key architectural component of 5G and future wireless networks. Offloading computational tasks from mobile devices to edge servers instead of the cloud results in internet bandwidth savings and circumvents the long delays involved in communicating the data load of the offloaded tasks to a cloud data centre residing somewhere on the internet. Above all, edge computing augments the compute and memory limitations of edge devices.
Jaya Prakash Champati and Ben Liang. Single Restart with Time Stamps for Parallel Task Processing with Known and Unknown Processors. IEEE Transactions on Parallel and Distributed Systems. 31 - 1, pp. 187 - 200. 2020.
Jaya Prakash Champati and Ben Liang. Semi-Online Algorithms for Computational Task Offloading with Communication Delay. IEEE Transactions on Parallel and Distributed Systems. 28 - 4, pp. 1189 - 1201. 2017.