主に3つのトピックに関する研究を行っています。
(1)記号創発ロボティクス・ロボット学習:マルチモーダルなセマンティックマップ(場所概念)の形成とナビゲーションなどのサービスタスクへの活用
(2)脳参照アーキテクチャ駆動開発:神経科学的知見に基づく確率的生成モデルの設計と実証
(3)認知発達システム・記号創発システム:言語獲得・創発コミュニケーション・集合的予測符号化の数理モデル化とエージェントシミュレーション
Akira Taniguchi, Yoshinobu Hagiwara, Tadahiro Taniguchi, and Tetsunari Inamura, "Online Spatial Concept and Lexical Acquisition with Simultaneous Localization and Mapping", IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS2017), pp.811-818, Sep 2017. in Vancouver, Canada. DOI: 10.1109/IROS.2017.8202243 [Acceptance Rate: 45%] [IEEE Xplore] [arXiv] [video] [Slide]
We propose an online learning algorithm based on a Rao-Blackwellized particle filter for spatial concept acquisition and mapping. We have proposed a nonparametric Bayesian spatial concept acquisition model (SpCoA). We propose a novel method (SpCoSLAM) integrating SpCoA and FastSLAM in the theoretical framework of the Bayesian generative model. The proposed method can simultaneously learn place categories and lexicons while incrementally generating an environmental map. Furthermore, the proposed method has scene image features and a language model added to SpCoA. In the experiments, we tested online learning of spatial concepts and environmental maps in a novel environment of which the robot did not have a map. Then, we evaluated the results of online learning of spatial concepts and lexical acquisition. The experimental results demonstrated that the robot was able to more accurately learn the relationships between words and the place in the environmental map incrementally by using the proposed method.
Tomochika Ishikawa, Akira Taniguchi, Yoshinobu Hagiwara, Tadahiro Taniguchi, "Active Semantic Mapping for Household Robots: Rapid Indoor Adaptation and Reduced User Burden", IEEE International Conference on Systems, Man, and Cybernetics (SMC), Oct. 2023. in Hawaii, USA. [Acceptance Rate: 55.61% ] [Project page] [LINK]
SpCoAE:
Akira Taniguchi, Yoshiki Tabuchi, Tomochika Ishikawa, Lotfi El Hafi, Yoshinobu Hagiwara, Tadahiro Taniguchi, "Active Exploration based on Information Gain by Particle Filter for Efficient Spatial Concept Formation", Advanced Robotics, Vol. 37, No. 13, pp.840-870 (31 pages), 2023. DOI: 10.1080/01691864.2023.2225175 [LINK] [arXiv]
Active semantic mapping is essential for service robots to quickly capture both the environment’s map and its spatial meaning, while also minimizing users’ burdens during robot operation and data collection. SpCoSLAM, which is a semantic mapping with place categorization and simultaneous localization and mapping (SLAM), offers the advantage of not being limited to predefined labels, making it well-suited for environmental adaptation. However, SpCoSLAM presents two issues that increase users’ burdens: 1) users struggle to efficiently determine a destination for the robot’s quick adaptation, and 2) providing instructions to the robot becomes repetitive and cumbersome. To address these challenges, we propose Active-SpCoSLAM, which enables the robot to actively explore uncharted areas while employing CLIP as image captioning to provide a flexible vocabulary that replaces human instructions. The robot determines its actions by calculating information gain integrated from both semantics and SLAM uncertainties.
Tidy-up tasks by service robots in home environments are challenging in robotics applications because they involve various interactions with the environment. In particular, robots are required not only to grasp, move, and release various home objects but also to plan the order and positions for placing the objects. We propose a novel planning method that can efficiently estimate the order and positions of the objects to be tidied up by learning the parameters of a probabilistic generative model. The model allows a robot to learn the distributions of the co-occurrence probability of the objects and places to tidy up using the multimodal sensor information collected in a tidied environment. Additionally, we develop an autonomous robotic system to perform the tidy-up operation. We evaluate the effectiveness of the proposed method by an experimental simulation that reproduces the conditions of the Tidy Up Here task of the World Robot Summit 2018 international robotics competition. The simulation results show that the proposed method enables the robot to successively tidy up several objects and achieves the best task score among the considered baseline tidy-up methods.
Akira Taniguchi, Tadahiro Taniguchi, and Angelo Cangelosi, "Cross-Situational Learning with Bayesian Generative Models for Multimodal Category and Word Learning in Robots", Frontiers in Neurorobotics, Vol. 11, No. 66, 19 pages, 2017. DOI: 10.3389/fnbot.2017.00066 [LINK]
Human infants can acquire word meanings by estimating the relationships among multimodal information and words. In this paper, we propose a novel Bayesian generative model that can form multiple categories based on each sensory-channel and can associate words with any of four sensory-channels (action, position, object, and color). This paper focuses on cross-situational learning using the co-occurrence between words and information of sensory-channels in complex situations.
We conducted a learning experiment using a simulator and a real humanoid iCub robot. In the experiments, a human tutor provided a sentence that describes an object of visual attention and an accompanying action to the robot. The experimental results showed that the proposed method was able to estimate the multiple categorizations and to learn the relationships between multiple sensory-channels and words accurately. In addition, we conducted an action generation task and an action description task based on word meanings learned in the cross-situational learning experiment. The experimental results showed the robot could successfully use the word meanings learned by using the proposed method.
In building artificial intelligence (AI) agents, referring to how brains function in real environments can accelerate development by reducing the design space. In this study, we propose a probabilistic generative model (PGM) for navigation in uncertain environments by integrating the neuroscientific knowledge of hippocampal formation (HF) and the engineering knowledge in robotics and AI, namely, simultaneous localization and mapping (SLAM). We follow the approach of brain reference architecture (BRA) (Yamakawa, 2021) to compose the PGM and outline how to verify the model. To this end, we survey and discuss the relationship between the HF findings and SLAM models. The proposed hippocampal formation-inspired probabilistic generative model (HF-PGM) is designed to be highly consistent with the anatomical structure and functions of the HF. By referencing the brain, we elaborate on the importance of integration of egocentric/allocentric information from the entorhinal cortex to the hippocampus and the use of discrete-event queues.
Spatial cognition in hippocampal formation (HF) is believed to be key to developing robot self-localization. Theta phase precession is thought to discretize experiences and compress the representation of movement trajectories within a timestep. External stimuli are sampled within theta-wave cycles. It is believed to be encoded by past, current, and future events in the phase of theta waves [Terada et al., 2017].
We propose a self-localization approach, DEQ-MCL, based on the discrete event queue hypothesis associated with theta phase precession within the HF. Our method estimates the posterior distribution of states representing a queue, including past, present, and future. This approach enables the smoothing of the posterior distribution of past states using current observations and the weighting of the joint distribution by considering the feasibility of future states. Our results suggest that the proposed method could improve the self-localization performance in indoor environments.
Takeshi Nakashima, Shunsuke Otake, Akira Taniguchi, Katsuyoshi Maeyama, Lotfi El Hafi, Tadahiro Taniguchi, Hiroshi Yamakawa, "Hippocampal formation-inspired global self-localization: quick recovery from the kidnapped robot problem from an egocentric perspective", Frontiers in Computational Neuroscience, Vol. 18, May 2024. DOI: 10.3389/fncom.2024.1398851 [LINK] [Project page]
It remains difficult for mobile robots to continue accurate self-localization when they are suddenly teleported to a location different from their beliefs during navigation. By incorporating insights from neuroscience into developing a spatial cognition model for mobile robots, it is possible to acquire the ability to respond appropriately to changing situations, similar to living organisms. Indeed, recent neuroscience research has shown that during teleportation in rat navigation, neural populations of place cells in the CA-3 region of the hippocampus, which are sparse representations of each other, switch discretely.
The spatial cognition model was realized by integrating the recurrent state–space model, a world model, with Monte Carlo localization to infer allocentric self-positions within the framework of neuro-symbol emergence in the robotics toolkit. The spatial cognition model, which models the CA-1 and -3 regions with each latent variable, demonstrated improved self-localization performance of mobile robots during teleportation in a simulation environment. Moreover, it was confirmed that sparse neural activity could be obtained for the latent variables corresponding to CA-3. These results suggest that spatial cognition models that incorporate neuroscience insights can contribute to improving the self-localization technology for mobile robots.