Vehicle Technology Directorate, ARLApproved for public release; distribution unlimited. ii REPORT DOCUMENTATION PAGE Form Approved OMB No. 0704-0188Public reporting burden for this collection of information is estimated to average 1 hour per response, including the time for reviewing instructions, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing the collection information. Send comments regarding this burden estimate or any other aspect of this collection of information, including suggestions for reducing the burden, ARL-TR-4800 SPONSOR/MONITOR'S ACRONYM(S) 9. SPONSORING/MONITORING AGENCY NAME(S) AND ADDRESS(ES) SPONSOR/MONITOR'S REPORT NUMBER(S) DISTRIBUTION/AVAILABILITY STATEMENTApproved for public release; distribution unlimited. SUPPLEMENTARY NOTES ABSTRACTIn theory, autonomous robotic swarms can be used for critical Army tasks (i.e., accompanying convoys); however, the Soldier controlling the swarm must be able to monitor swarm status and correct actions, especially in disrupted or degraded conditions. For this two-year Director's Research Initiative (DRI), we designed metacognition algorithms and Soldier-swarm display concepts to allow Soldiers to efficiently interact with a robotic swarm participating in a representative convoy mission. We used a potential field approach for swarm control because it scales easily to large heterogeneous swarms and allows users to dynamically alter swarm behavior by adjusting field parameters. The Soldier-swarm interface displayed swarm and convoy geospatial position; swarm health and communication; and convoy status information, using visual, auditory, and tactile combinations. We measured swarm metacognition by determining the proportion of time the simulated swarm could maintain a specific orbital ring around the convoy over six terrains in 13-min scenarios. We tested interface effectiveness in a laboratory study using 16 male Marines (volunteers) with a mean age of 19 years. The metacognition results showed that the swarm could maintain the pre-defined dispersion more than 85% of the time in each terrain. Using multimodal displays, Soldier workload decreased and performance increased (i.e., response time reduced).
Abstract-In theory, autonomous robotic swarms can be used for critical Army tasks, including accompanying vehicle convoys to provide security and enhance situational awareness. However, the Soldier providing swarm supervisory control must be able to correct swarm actions, especially in disrupted or degraded conditions. Dynamic map displays are visual interfaces that can be useful for swarm supervisory control tasks, because they can show the spatial positions of objects of interest (e.g., people, robots, swarm members, and vehicles), at different locations (e.g., on roads and intersections), while allowing user commands as well as world changes, often in real time. In this study, multimodal speech and touch controls were designed for a U.S. Army Research Laboratory dynamic map display to allow users to provide supervisory control of a simulated robotic swarm. This experiment explored the use of sequential multimodal touch and speech commands for placement of swarm-related map objects at different map locations. The criterion variable was temporal binding, the time between the onset of each command in the sequence, relative to the system's ability to fuse the two sequential commands into a unitary response. User preference of modality for the first command was also measured. These concepts were tested in a laboratory study using 12 male Marine volunteers with a mean age of 19 years. Results indicated significant differences in temporal binding for different map objects and map locations. Additionally, nine out of 12 Marines used speech commands approximately 75% or more of the time, while the remaining three Marines used touch commands first approximately 75% or more of the time. Temporal binding was significantly shorter for touch-first than for speech-first commands.Suggestions for future research and future applications to robotic command and control systems are described.
Public reporting burden for this collection of information is estimated to average 1 hour per response, including the time for reviewing instructions, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing the collection information. Send comments regarding this burden estimate or any other aspect of this collection of information, including suggestions for reducing the burden, to Department of Defense, Washington Headquarters Services, Directorate for Information Operations and Reports (0704-0188), 1215 Jefferson Davis Highway, Suite 1204, Arlington, VA 22202-4302. Respondents should be aware that notwithstanding any other provision of law, no person shall be subject to any penalty for failing to comply with a collection of information if it does not display a currently valid OMB control number. PLEASE DO NOT RETURN YOUR FORM TO THE ABOVE ADDRESS. REPORT DATE (DD-MM-YYYY)December 2005 ARL-TR-3632 SPONSOR/MONITOR'S ACRONYM(S) 9. SPONSORING/MONITORING AGENCY NAME(S) AND ADDRESS(ES) SPONSOR/MONITOR'S REPORT NUMBER(S) DISTRIBUTION/AVAILABILITY STATEMENTApproved for public release; distribution is unlimited. SUPPLEMENTARY NOTES ABSTRACTMission demands have made the robotics collaboration operator control unit (OCU) into a relatively dynamic, demanding, cognitively complex system where Soldiers must perform multiple tasks such as controlling multiple robots and processing large amounts of information in environments that sometimes contain high levels of noise. Research and modeling data indicate that audio display technologies would be very useful in OCU applications such as guiding visual display search. The purpose of this study was to examine the effectiveness of the integration of auditory display technologies in visual search tasks such as those that occur in robotic OCUs. Independent variables were audio signal mapping scheme, type of verbal positional cue, and visual target azimuth. Dependent variables were visual target search time and the National Aeronautics and Space Administration Task Load Index workload rating of the target search task. Participants were 36 students (15 males and 21 females) from Harford Community College. The results indicated that the use of auditory signal mapping and verbal positional cues significantly reduced visual display search time and workload and that positional cues mixed with specific audio mappings were the most efficient means of reducing search time. Specific design recommendations are made regarding the use of auditory signals in environments with narrow field-of-view visual displays. SUBJECT TERMSauditory assisted visual search; auditory displays SECURITY
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2025 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.