Abstract: "The talk explores the role of humans in artistic co-creation with AI, focusing on music and dance. Moving beyond the view of humans as just end-users, we frame co-creation as an iterative process where artists, audiences, and AI systems jointly shape the creative outcome. Through case studies in choreographic generation and sample-based music composition, we show how human knowledge (expressed through constraints, prompt design, and evaluation) consistently improves both the quality and acceptance of AI-generated works. We present protocols for evaluation, methods for embedding domain expertise into AI-driven creativity, and patterns of human intervention across stages of the generative process."
"Leveraging the multi-modal nature of communication in immersive XR environments"
Abstract: Communication is inherently multimodal: speech leans on gaze, pointing, and shared scene context. Unlike prior paradigms, spatial computing is uniquely positioned to exploit this richness. This keynote surveys a line of work that illustrate how blending explicit speech with behaviours and environmental signals makes mixed-reality collaboration and interaction more natural, efficient, and robust. We show how fusing speech with gaze, pointing, and semantic scene metadata resolves deictic references and yields more faithful interpretations of collaborative activity, we introduce EmBARDiment, an embodied agent that turns gaze-driven contextual memory and voice into low-friction, implicit assistance for multi-window workflows finally we revisit “Put That There” by coupling LLMs with head-gaze, pointing, and scene understanding to enable intent-based placement and manipulation of virtual objects."