o Paper 1: Forensic Analysis of Indirect Prompt Injection Attacks on LLM Agents – Maxim Chernyshev (Deakin University, Australia), Zubair Baig (Deakin University, Australia), and Robin Doss (Deakin University, Australia)
o Paper 2: Pitfalls of Generic Large Language Models (GLLMs) from Reliability and Security Perspectives – Dipankar Dasgupta (The University of Memphis, USA) and Arunava Roy (The University of Memphis, USA)
o Paper 3: Large Language Models for Hardware Security – Hammond Pearce (University of New South Wales, Australia) and Benjamin Tan (University of Calgary, Canada)
o Paper 4: Secure Lightweight Computation for Federated N-Gram Language Models – Tho Thi Ngoc Le (HUTECH University, Vietnam) and Tran Viet Xuan Phuong (University of Arkansas at Little Rock, USA)
o Paper 5: Probing Robustness of In-context Learning in LLM Classification Predictions Under Label Noise – Sriya Ayachitula (Ardsley School, USA), Chinmay Kundu (KIIT University, India), and Birendra Mishra (University of California, Riverside, USA)