4th Workshop on
Semantic Policy and Action Representations for Autonomous Robots (SPAR)
November 8th, 2019 - Macau, China
Room: LG-R09
Room: LG-R09
It has been a long-standing question whether robots can reach human level of intelligence that understands the essence of observed actions and imitates them even under different circumstances. Contemporary research in robotics and machine learning has attempted to solve this question from two different perspectives: One in a bottom-up manner by, for instance, solely relying on perceived continuous sensory data, whereas the other by approaching rather from the symbolic level in a top-down fashion. Although there have been shown encouraging results in both flows, understanding and imitation of actions have yet to be fully solved.
Action semantics stands as a potential glue for bridging the gap between a symbolic action representation and its corresponding continuous signal level description. Semantic representation provides a tool for capturing the essence of action by revealing the inherent characteristics. Thus, semantic features help robots to understand, learn, and generate policies to imitate actions even in various styles with different objects. Thus, more descriptive semantics yields robots with greater capability and autonomy. In this full-day workshop, we aim at answering two major questions.
This workshop focuses on new technologies that allow robots to learn generic semantic models for different tasks. In this workshop, we will bring together researchers from diverse fields, including robotics, computer vision, and machine learning in order to overview the most recent scientific achievements and the next break-through topics, and also to propose new directions in the field.
The DFG Collaborative Research Center 1320: Everyday Activity Science and Engineering - EASE (https://ease-crc.org/)
US NSF #1750082: Visual Recognition with Knowledge - VR-K
Vinnova FFI project SHARPEN, under grant agreement no. 2018-05001