Program

1st day (March 4th)

9:00-9:15 Opening

9:15-10:00 Keynote 1 (Giuseppe Riccardi, University of Trento Italy)
1. Conversational AI to Benefit Individuals

Research in human-machine dialogue (aka conversational AI) has been driven by the quest for open-domain, knowledgeable and multimodal agents. In contrast,  the complex problem of designing , training and evaluating a conversational system and its components is currently reduced to a) prompting large language models, b) coarse  evaluation of machine responses and c) poor management of the affective signals. In this talk we will review the current state-of-the-art in human-machine dialogue research and its limitations. We will present the most challenging frontiers of conversational AI when the objective is to create personal conversational systems that benefit individuals. In this context we will report experiments and RCT trials  of so-called personal healthcare agents supporting individuals and healthcare professionals.

10:00-10:30 coffee break

10:30-12:10 Oral Session 1 (Cognitive and Psychological perspective of dialogue systems)

12:10-13:25 Lunch (lunch box will be provided)

13:25-13:30 LT of online posters

13:30-15:00 Poster/Demo/Position Session

15:00-15:30 coffee break

15:30-16:30 Oral Session 2 (Dialogue and LLMs)

16:50-18:20 Poster/Demo/Position Session 2 

one language?; Jacqueline Brixey (University of Southern California)*; David Traum (USC Institute for Creative Technologies)


2nd day (March 5th)

9:00-9:05 Announcement

9:05-9:50 Keynote 2 (Tatsuya Kawahara, Kyoto University)
28. Semi-autonomous Dialogue for Cybernetic Avatars and Ainu Speech Processing

Spoken dialogue systems (SDS) have made dramatic advances thanks to improved speech technology and large language models. However, they still have limitations in terms of human-level empathy. Purely autonomous systems might be out of control and produce unexpected responses. Meanwhile, avatars have become prevailing in online communications since the pandemic.  A hybrid of avatars with AI and robotics, called cybernetic avatars (CA), is expected to complement each other and provide human-level services to many people in parallel and simultaneously.  The semi-autonomous dialogue system can be implemented for either a robot or a CG-based avatar.  The talk addresses this new project of spoken dialogue systems for cybernetic avatars, which is sponsored by the Moonshot Research and Development Program in Japan.

The talk also introduces the projects on speech processing of the Ainu language, a critically endangered language in Hokkaido.  Specifically, automatic speech recognition is developed for annotation of the folklore archive, and speech synthesis is explored for language learning.

9:50-10:10 coffee break

10:10-11:30 Oral Session 3 (Best Paper Session)

11:30-12:30 Sponsor Lunch (lunch box will be provided)

12:30-19:00 Excursion (Upopoy; National Ainu Museum)

19:00-21:00 Banquet (Hotel)

3rd day (March 6th)

9:00-9:15 Announcement

9:15-10:00 Keynote 3 (Yukiko Nakano, Seikei University)
33. Modeling Multimodal Interactions to Enhance the User Understanding in Dialogue Systems

In Human-Agent interaction using virtual characters and robots, both the user and the conversational agent display nonverbal signals. In order to achieve the system’s accurate understanding of the content and state of such multimodal interactions, it is essential to enhance the agent’s ability to properly interpret nonverbal information such as prosody, facial expressions, and eye gaze, in addition to verbal information. This talk will discuss our research in multimodal interaction, addressing issues for detecting important events and utterances that provides richer representation of dialogue states and estimating user characteristics that would be useful in understanding group dynamics in multiparty communications. This talk will cover methodologies of interaction analysis, multimodal machine learning, multimodal corpus collection, and communicative behavior annotation, and will show how multimodal information contributes to improving the performance of machine learning models.  

10:00-10:30 coffee break

10:30-12:10 Oral Session 4 (Dialogue Related Technologies)

12:10-13:30 Lunch (lunch box will be provided)

13:30-14:50 Oral Session 5 (Dialogue Evaluation)

14:50-15:20 Coffee Break

15:20-16:20 Panel
Topic: Research directions and sustainability for linguistic diversity in the age of LLMs?

Panelists:
•Yun-Nung (Vivian) Chen (National Taiwan University)
•Luis Fernando D'Haro (Universidad Politécnica de Madrid)
•Jacqueline Brixey (University of Southern California)

16:30-17:00 Closing (Includes introduction of the Best Paper Award)

Instruction to presenters

Oral presentations

Any oral presentations have 15 minutes for presentation and 5 minutes for QAs. Presentations will be distributed to online participants via Zoom. We will record the video to share with participants for a short period of time after the workshop. If you or your organization requires us not to record the video, please get in touch with organizers <iwsds2024@gmail.com>.

If you plan to present online, please share your screen after the chair requires you. 

Poster/demo/position presentations

Any poster/demo/position presentation has a poster board suitable for an A0 portrait. If you need a power supply, please get in touch with organizers <iwsds2024@gmail.com>. Other equipment, such as monitors, will not be provided.

If you plan to present online, please send your poster and 3-minute video to organizers by February 23rd. We will play your video at the LT session and print/hang your poster on your poster board.