Publication:
Multi-modal data fusion for people perception in the social robot haru

Loading...
Thumbnail Image

Publication date

Reading date

Event date

Start date of the public exhibition period

End date of the public exhibition period

Advisors

Authors of photography

Person who provides the photography

Journal Title

Journal ISSN

Volume Title

Publisher

Springer Nature
Export

Research Projects

Organizational Units

Journal Issue

Abstract

This article presents a people perception software architecture and its implementation, focused on the information of interest from the point of view of a social robot. The key modules employed to get the different people features, such as the body parts location, the face and hands information, and the speech, from a set of possible devices and configurations are described. The association and combination of these features using a temporal and geometric fusion system are explained in detail. A high-level interface for Human-Robot interaction using the resulting information from the fused people is proposed. The paper presents experimental results evaluating the relevant aspects of the system.

Doctoral program

Related publication

Research projects

Description

This work is partially supported by Programa Operativo FEDER Andalucia 2014-2020, Consejeria de Economía y Conocimiento (DeepBot, PY20\_00817) and the project PLEC2021-007868, funded by MCIN/AEI/10.13039/501100011033 and the European Union NextGenerationEU/PRTR.
Proyectos de investigación Proyecto DeepBot (PY20_00817) Proyecto NHoA (PLEC2021-007868)

Bibliographic reference

R Ragel, R Rey, Á Páez, J Ponce, K Nakamura, F Caballero, L Merino and R Gómez. "Multi-modal Data Fusion for People Perception in the Social Robot Haru", International Conference on Social Robotics, 174-187, 2022

Photography rights