Presentation + Paper
6 May 2021 Designing a mixed reality interface for autonomous robot-based change detection
Author Affiliations +
Abstract
Robots, equipped with powerful modern sensors and perception algorithms, have enormous potential to use what they perceive to provide enhanced situational awareness to their human teammates. One such type of information is changes that the robot detects in the environment that have occurred since a previous observation. A major challenge for sharing this information from the robot to the human is the interface. This includes how to properly aggregate change detection data, present it succinctly for the human to interpret, and allow the human to interact with the detected changes, e.g., to label, discard, or even to task the robot to investigate, for the purposes of enhanced situational awareness and decision making. In this work we address this challenge through the design of an augmented reality interface for aggregating, displaying, and interacting with changes detected by an autonomous robot teammate. We believe the outcomes of this work could have significant applications to Soldiers interacting with any type of high-volume, autonomously-generated information in Multi-Domain Operations.
Conference Presentation
© (2021) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Christopher Reardon, Jason Gregory, Carlos Nieto-Granda, and John G. Rogers III "Designing a mixed reality interface for autonomous robot-based change detection", Proc. SPIE 11759, Virtual, Augmented, and Mixed Reality (XR) Technology for Multi-Domain Operations II, 117590J (6 May 2021); https://doi.org/10.1117/12.2587494
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Robots

Mixed reality

Situational awareness sensors

Augmented reality

Dismounted soldiers

Environmental sensing

Information operations

Back to Top