Attempting to Aggregate Perceptual Constructs From Deep Neural Networks for Video and Audio Interaction Representation
- Resource Type
- Conference
- Authors
- Maheux, Marc-Antoine; Auclair, Guillaume; Warren, Philippe; Letourneau, Dominic; Michaud, Francois
- Source
- 2023 32nd IEEE International Conference on Robot and Human Interactive Communication (RO-MAN) Robot and Human Interactive Communication (RO-MAN), 2023 32nd IEEE International Conference on. :1302-1307 Aug, 2023
- Subject
- Communication, Networking and Broadcast Technologies
Computing and Processing
Robotics and Control Systems
Signal Processing and Analysis
Visualization
Aggregates
Human-robot interaction
Artificial neural networks
Assistive robots
Older adults
Robots
- Language
- ISSN
- 1944-9437
Socially Assistive Robots are foreseen as having the potential to improve the quality of life of older adults and individuals with mental disabilities. Natural human-robot interaction in everyday settings may require robots that are capable of understanding what is happening in their operating environments so that they can respond appropriately to the experienced situations and engage people in meaningful ways. This paper presents an approach using perceptual constructs to represent what is being observed by the robot. Perceptual constructs are derived from deep neural networks used to process visual and audio data. The objective is to derive a compressed representation of the interactions observed by the robot in real-life settings. Results are provided from observations made by a robot of a room with human activity over a two-week period, outlining what works and remaining challenges.