VAM-HRI 2024 — 2024 International Workshop on Virtual, Augmented, and Mixed-Reality for Human-Robot Interactions at HRI 2024, 2024

Prototyping Mid-Air Display for Anywhere Robot Communication With Projected Spatial AR

Uthman Tijani and Zhao Han

Prototyping Mid Air Display

Abstract

Augmented reality (AR) allows visualizations to be situated where they are relevant, e.g., in a robot’s operating environment or task space. Yet, headset-based AR suffers a scalability issue because every viewer must wear a headset. Projector-based spatial AR solves this problem by projecting augmentations onto the scene, e.g., recognized objects or navigation paths viewable to crowds. However, this solution mostly requires vertical flat surfaces that may not exist in open areas like warehouses, construction sites, and search and rescue scenes. Moreover, when humans are not co-located with the robot or situated at a distance, the projection may not be visible to humans. Thus, there is a need to create a flat, viewable surface for humans in such scenarios.

In this work, we designed a prototype of a fog screen device to create a flat, projectable display surface in the air. A robot with such a device mounted empowers it to communicate with humans in environments without flat surfaces or with only irregular surfaces in search and rescue sites. Specifically, we detail the design, covering underlying principles, material selection, 3D modeling, and evaluation. We will verify and iterate the design, explore the optimal mounting position, and conduct human evaluation in the future.


CCS Concepts

• Hardware → Displays and imagers; • Human-centered computing → Mixed / augmented reality; • Computer systems organization → Robotics.

Keywords

Fog screen, mid-air display, robot communication, projected AR, augmented reality (AR), human-robot interaction

Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the owner/author(s). VAM-HRI ’24, March 11, 2024, Boulder, CO, USA © 2024 Copyright held by the owner/author(s).
(HRI)

ACM Reference Format:

Uthman Tijani and Zhao Han. 2024. Prototyping Mid-Air Display for Anywhere Robot Communication With Projected Spatial AR . In Proceedings of the 7th International Workshop on Virtual, Augmented, and Mixed-Reality for Human-Robot Interactions (VAM-HRI ’24), March 11, 2024, Boulder, CO, USA. 19 pages.


1 Introduction

In HRI, various communication modalities have been explored [25], including speech and audio [1, 18], gaze [11], visual displays [23], and body language like gestures and postures [8]. Research efforts have started to leverage augmented reality (AR) to enhance robot communication of non-verbal cues [35, 7, 12, 17, 27], including projected AR works [2, 7] where a robot’s navigation paths and manipulation intent were projected onto the ground and table surface.

concept
Figure 1: Proposed conceptual setup of our system, a low-fidelity representation of our plan to place the fog screen device and the projector on the robot. Details of the fog screen device are in the next pages.

Particularly, AR allows visual overlays situated in a robot’s task environment to externalize a robot’s internal states. Projector-based AR further improves AR as it offers scalability such that a crowd of people no longer need to wear AR headsets. However, it is crucial to consider different environments, e.g., settings that lack projectable surfaces or those not viewable by humans, instances where humans are not co-located with or further away from the robot, and situations where physical obstructions may affect the visibility and effectiveness of AR visualizations. While these work [25, 7, 12, 17, 27] improve the understanding of a robot’s intent and behaviors like manipulation and navigation using AR, we must solve the problem that some environments lack projectable surfaces.

To retain the benefits of projected AR, we must answer this question: “how can we leverage this technology for robot communication when there is no available or suitable (e.g., irregular) surface to project onto?” In response, we propose using a fog screen device to create a mid-air flat display medium for robots to communicate in environments lacking projectable surfaces. This is particularly beneficial for open environments like warehouses, construction sites where suitable surfaces may be scarce or search and rescue scenes with irregular surfaces.

Specifically, we present a prototype design of a flat fog screen device. Currently, we are building the first version of the device and verifying the optimal mounting position of the fog screen device, together with a projector, on a specific robot platform—the Fetch mobile manipulator. Our initial concept plan (Figure 1) is to place the fog screen device in the front right of the robot’s base and place the projector in the front left of the base, therefore making it possible to project content onto the generated fog screen.

In the subsequent sections, we delve into the specific design considerations–how the fog screen device works–and evaluation plans to verify the proposed fog screen device with humans. Through this work, we strive to contribute practical insights and address communication challenges in dynamic environments.

2 Related Work

2.1 Projected Augmented Reality (AR)

This section explores how projected AR enhances human-robot communication in various aspects, from conveying robot intentions to addressing challenges in understanding robot navigation in populated areas. For example, Chadalavada et al. [2] aimed to improve communication between humans and robots in collaborative workspaces. By projecting a robot’s internal states onto the shared floor space, the paper investigates how AR can convey a robot’s navigation intent. Han et al. [7] also tried to address the challenge of understanding the navigation intent of mobile robots in populated areas while also open-sourcing the hardware and software implementation for the projected AR technology. The emphasis was on directional projections, specifically arrow projections, as visual cues to convey the robot’s intended navigation path. The goal was to provide a practical solution for visualizing robot navigation intent, contributing to improved participant perception, enhanced comfort, and behavioral changes in response to the projections. Our work also belongs to the projected AR family, but we focus on creating a fog screen for robot communication.

2.2 Mid-Air Display

The exploration of mid-air display technologies has led to diverse innovations addressing challenges in cost [1013], bulkiness and low resolution [13], fog flow deformation [14], tactile feedback [2021], and mounting position [6].

Rakkolainen et al. [16] explores the fog screen’s potential application domains like entertainment, museums, and trade shows. They outlined future research directions to address challenges related to high costs, multi-modality interactions, and tracking using new sensors. Particularly related to robotics, Scheible et al. [22] introduced “Displaydrone”, a system that has a drone, a Pico projector, and a smartphone to create a flying interactive display for projecting content onto walls and objects in physical space, such as building facades and rock surfaces. This work highlights challenges in mobile interactive displays, such as short battery life, legal issues, projection viewing distance, and stable hovering. Finally, Otao and Koga [14] addressed the difficulty of detecting the deformation of fog flow caused by a user’s action in real-time by proposing a feed-forward approach to create pseudo-synchronized image contents along with the deformed fog.

2.3 Feedback for Mid-Air Display Interactions

To make user interaction with fog screen display responsive, Sand et al. [2021] investigated the effect of wearable vibrotactile feedback devices for mid-air gestural interaction with large fog screens on user preference [21]. Two finger-based gestures, tapping and dwell-based (i.e., the user’s finger stays longer in the mid-air display) were compared. A user study showed that tapping without haptic feedback was the most preferred gesture type, while haptic feedback was preferred for the dwell-based gesture. Similarly, Dancu et al. [6] investigated optimal wearable mid-air display placement during map navigation, comparing wrist and chest mounting approaches. Results from the user study showed that wrist-mounting is considered safer and preferable for map navigation.

2.4 Reconfigurable Mid-Air 3D Display

Researchers also attempted to display volumetric 3D content by creating 3D fog displays. Lam et al. [10] used columns of upward-flowing laminar fog which is reconfigurable (fog particles can be re-positioned), allowing for true 3D perception without AR headsets, as projected images scatter at different depths. Lam et al. [9] later improved their work [10] by achieving a higher resolution, enabling users to touch and manipulate virtual objects with wide viewing angles directly. Tokuda et al. [24] also introduced combining shape-changing interfaces and mid-air fog displays, which can adapt their shape to support 2D and 3D visual content.

2.5 Compact Cost-Effective Mid-Air Display

While most works focused on user interaction [142021], shape-changing, reconfigurable fog screens and volumetric 3D content display [91024], Norasikin et al. [13] addressed the issue of high cost, bulkiness, and inability to provide higher resolution of mid-air displays. SonicSpray [13] addressed these challenges by offering a compact, low-cost solution that can precisely control laminar fog (suspension of tiny solid or liquid particles in gas) in mid-air. As some of the works [152021] focused on tactile feedback for fog screen displays, our focus is mainly on creating compact fog screen displays for robot communication. So far, fog screen creation has not been explored for communication in the HRI community.

3 Design

Our goal is to design a compact, portable fog screen device that can be mounted conveniently on a Fetch robot to be projected onto by a projector, as seen in Figure 1.

top view annotation
Figure 2: Top view of our fog screen device showing its internal mechanism. The blue particles within the device generated by the fog machine fill the middle container and exit the right opening. The laminar airflow former (Figure 3) regulates the dispersed fog into a laminar flow to form a screen.

As seen in Figure 2, the device begins by making water into tiny droplets to generate an uncontrolled mist or fog, then uses a set of fans to generate high-pressure air, and finally uses the laminar airflow formers to transform the fog into a stable and flat fog screen.

First, we used an off-the-shelf fog machine to generate fog with a special liquid called “fog juice” [26]. It consists of water, glycerin, and propylene glycol. When this liquid is heated, it turns into mist or fog. A built-in pump pushes this mist through a small opening on the right. When it meets the cooler air outside, it turns into opaque fog that can be seen. Although the fog can be produced by an ultrasonic atomizer placed in a container of water, we realized that using the fog machine is cleaner (no concerns for water leakage), flexible (changeable device orientation), and produces denser fog using fog juice (mixtures of water and glycerin) [26].

The middle fog container will host the fog machine. In the container, the generated fog will accumulate first before coming out of the thin opening on the right. Because the generated fog at this stage is dispersed as it comes out, it is unsuitable to project onto as it does not make it a flat screen. To solve this, it must be controlled to follow a laminar flow resembling a thin layer of surface in mid-air for projection.

airflow former
Figure 3: 3D CAD model of laminar airflow former, passage for high-pressure air from the fans to make the fog exit in a particular direction to form a flat screen. Taking a closer look at the airflow former design, it is similar to gluing a series of plastic straws together by the side or like a honeycomb.

The set of fans at both sides (Figure 2 top and bottom) and the airflow formers (Figure 2 right and Figure 3) are designed to achieve this. The fans generate high-pressure air by sucking in air through the vents at the left end of the device and traveling right to the airflow formers, where the airflow is made laminar. These laminar high-pressure air on both sides serve as a barrier for the fog, forcing it to remain within it, creating a flat, thin layer of fog in mid-air suitable for projectors to project onto.

side view annotation
Figure 4: Full side view of our fog screen device with fans. This is how it will be positioned on the robot, allowing a wide viewing and projection angle.

Figure 4 shows a full side view of the device, giving a closer look at the smaller view in Figure 2.

3.1 Materials

housing
Figure 5: 3D CAD model of the case of the proposed fog screen device, about 640mm wide. (a) The housing is more like the device framing. It has slots for the airflow formers and the fog container. (b) This is how the device looks after all designed parts have been assembled (housing, airflow formers, and fog container).

We plan to test our prototype by first 3D-printing all the parts we have designed. Besides the laminar airflow former shown in Figure, we have also designed a 3D CAD model of the case using SOLIDWORKS, which is shown in Figure 5. The case houses and protects the internal components like the fans and external components like the airflow formers. This initial case design was inspired by Hoverlay II [28]. Additionally, we plan to buy the fans, the fog machine, and a portable battery to power the fans and the fog machine.

3.2 Placement on Robot

The housing design structure and the fog machine allow the device to be placed vertically as shown in Figure 1. We plan to place the fog screen device on a Fetch robot’s base, which has M5 holes with a grid spacing of 50𝑚𝑚 [19]. The projector will also be placed on the base near the fog screen device.

4 Evaluation Plan

First, we will build the prototype with all the parts 3D printed or bought and verify our theory to see if the device can produce a desirable flat fog screen. This may take several iterations. We will examine whether the airflow former can prevent the fog from dispersing and keep the fog flow laminar, i.e., keep the airflow former parallel to the fog emission points. A slight error in the angle will disperse the fog from a flat screen shape. We will also examine how wide the screen can be and adjust accordingly.

Our next step is to verify the responsiveness of the fog screen generation. We will measure the time the fog machine and the fog screen device take to begin generating fog.

Third, we plan to test reliability for consistently generating usable fog for projection use. Here, we determine the performance or efficiency change over extended periods of use. The changes in the fog density and how long the fog machine can keep producing dense and wide, i.e., usable, fog will be closely monitored.

Fourth, we plan to test the robustness of the generated fog in different wind conditions to make it suitable for outdoor use. We expect that the high-pressure wind generated by the airflow former helps achieve robustness.

Formal human evaluation would be the final phase, getting user perception and feedback for potential design guidelines and improvements. In the far future, we also plan to explore ways to provide privacy between the user and the robot by either stopping the projector or dispersing the fog in the presence of a third party.

5 Conclusion

We proposed a prototype design of a fog screen device to make a mid-air display surface to address the difficulty of robots leveraging scalable projected AR to communicate in environments lacking projectable surfaces. We plan to build and iterate the device and mount it onto a Fetch robot. In the future, we will further test its functionality, reliability, and user perception.

References

[1] Gabriele Bolano, Lawrence Iviani, Arne Roennau, and Ruediger Dillmann. 2021. Design and Evaluation of a Framework for Reciprocal Speech Interaction in Human-Robot Collaboration. In 2021 30th IEEE International Conference on Robot & Human Interactive Communication (RO-MAN). IEEE, 806–812.

[2] Ravi Teja Chadalavada, Henrik Andreasson, Robert Krug, and Achim J Lilienthal. 2015. That’s on my mind! robot to human intention communication through on-board projection on shared floor space. In 2015 European Conference on Mobile Robots (ECMR). IEEE, 1–6.

[3] Ravi Teja Chadalavada, Henrik Andreasson, Maike Schindler, Rainer Palm, and Achim J Lilienthal. 2020. Bi-directional navigation intent communication using spatial augmented reality and eye-tracking glasses for improved safety in human–robot interaction. Robotics and Computer-Integrated Manufacturing 61 (2020), 101830.

[4] Kishan Chandan, Vidisha Kudalkar, Xiang Li, and Shiqi Zhang. 2021. ARROCH: Augmented reality for robots collaborating with a human. In 2021 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 3787–3793.

[5] Michael D Coovert, Tiffany Lee, Ivan Shindev, and Yu Sun. 2014. Spatial augmented reality as a method for a mobile robot to communicate intended movement. Computers in Human Behavior 34 (2014), 241–248.

[6] Alexandru Dancu, Mickaël Fourgeaud, Mohammad Obaid, Morten Fjeld, and Niklas Elmqvist. 2015. Map navigation using a wearable mid-air display. In Proceedings of the 17th International Conference on Human-Computer Interaction with Mobile Devices and Services. 71–76.

[7] Zhao Han, Jenna Parrillo, Alexander Wilkinson, Holly A Yanco, and Tom Williams. 2022. Projecting robot navigation paths: Hardware and software for projected ar. In 2022 17th ACM/IEEE International Conference on Human-Robot Interaction (HRI). IEEE, 623–628.

[8] Heather Knight and Reid Simmons. 2016. Laban head-motions convey robot state: A call for robot body language. In 2016 IEEE international conference on robotics and automation (ICRA). IEEE, 2881–2888.

[9] Miu-Ling Lam, Bin Chen, and Yaozhung Huang. 2015. A novel volumetric display using fog emitter matrix. In 2015 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 4452–4457.

[10] Miu-Ling Lam, Bin Chen, Kit-Yung Lam, and Yaozhun Huang. 2014. 3D fog display using parallel linear motion platforms. In 2014 International Conference on Virtual Systems & Multimedia (VSMM). IEEE, 234–237.

[11] Songpo Li and Xiaoli Zhang. 2017. Implicit intention communication in human–robot interaction through visual behavior studies. IEEE Transactions on Human-Machine Systems 47, 4 (2017), 437–448.

[12] Rhys Newbury, Akansel Cosgun, Tysha Crowley-Davis, Wesley P Chan, Tom Drummond, and Elizabeth A Croft. 2022. Visualizing robot intent for object handovers with augmented reality. In 2022 31st IEEE International Conference on Robot and Human Interactive Communication (RO-MAN). IEEE, 1264–1270.

[13] Mohd Adili Norasikin, Diego Martinez-Plasencia, Gianluca Memoli, and Sriram Subramanian. 2019. SonicSpray: a technique to reconfigure permeable mid-air displays. In Proceedings of the 2019 ACM International Conference on Interactive Surfaces and Spaces. 113–122.

[14] Kazuki Otao and Takanori Koga. 2017. Mistflow: a fog display for visualization of adaptive shape-changing flow. In SIGGRAPH Asia 2017 Posters. 1–2.

[15] Karri Palovuori, Ismo Rakkolainen, and Antti Sand. 2014. Bidirectional touch interaction for immaterial displays. In Proceedings of the 18th International Academic MindTrek Conference: Media Business, Management, Content & Services. 74–76.

[16] Ismo Rakkolainen, Antti Sand, and Karri Palovuori. 2015. Midair user interfaces employing particle screens. IEEE computer graphics and applications 35, 2 (2015), 96–102.

[17] Christopher Reardon, Kevin Lee, John G Rogers, and Jonathan Fink. 2019. Communicating via augmented reality for human-robot teaming in field environments. In 2019 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR). IEEE, 94–101.

[18] Frederic Anthony Robinson, Oliver Bown, and Mari Velonaki. 2020. Implicit communication through distributed sound design: Exploring a new modality in human-robot interaction. In Companion of the 2020 ACM/IEEE International Conference on Human-Robot Interaction. 597–599.

[19] Fetch Robotics. 2024. Robot Hardware Overview — Fetch & Freight Research Edition Melodic documentation. https://docs.fetchrobotics.com/robot_hardware.html.

[20] Antti Sand, Ismo Rakkolainen, Poika Isokoski, Roope Raisamo, and Karri Palovuori. 2015. Light-weight immaterial particle displays with mid-air tactile feedback. In 2015 IEEE International Symposium on Haptic, Audio and Visual Environments and Games (HAVE). IEEE, 1–5.

[21] Antti Sand, Vera Remizova, I Scott MacKenzie, Oleg Spakov, Katariina Nieminen, Ismo Rakkolainen, Anneli Kylliäinen, Veikko Surakka, and Julia Kuosmanen. 2020. Tactile feedback on mid-air gestural interaction with a large fogscreen. In Proceedings of the 23rd International Conference on Academic Mindtrek. 161–164.

[22] Jürgen Scheible, Achim Hoth, Julian Saal, and Haifeng Su. 2013. Displaydrone: a flying robot based interactive display. In Proceedings of the 2nd ACM International Symposium on Pervasive Displays. 49–54.

[23] Elena Sibirtseva, Dimosthenis Kontogiorgos, Olov Nykvist, Hakan Karaoguz, Iolanda Leite, Joakim Gustafson, and Danica Kragic. 2018. A comparison of visualisation methods for disambiguating verbal requests in human-robot interaction. In 2018 27th IEEE international symposium on robot and human interactive communication (RO-MAN). IEEE, 43–50.

[24] Yutaka Tokuda, Mohd Adili Norasikin, Sriram Subramanian, and Diego Martinez Plasencia. 2017. MistForm: Adaptive shape changing fog screens. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems. 4383–4395.

[25] Christiana Tsiourti, Astrid Weiss, Katarzyna Wac, and Markus Vincze. 2017. Designing emotionally expressive robots: A comparative study on the perception of communication modalities. In Proceedings of the 5th international conference on human agent interaction. 213–222.

[26] Ed Vitz. 2008. Fog machines, vapors, and phase diagrams. Journal of Chemical Education 85, 10 (2008), 1385.

[27] Michael Walker, Hooman Hedayati, Jennifer Lee, and Daniel Szafir. 2018. Communicating robot motion intent with augmented reality. In Proceedings of the 2018 ACM/IEEE International Conference on Human-Robot Interaction. 316–324.

[28] Moritz Walter. 2014. Hoverlay II open hardware interactive midair screen. https://hackaday.io/project/205-hoverlay-ii.


Posted