Microsoft – Nweon Paper https://paper.nweon.com 映维网,影响力虚拟现实(VR)、增强现实(AR)产业信息数据平台 Tue, 26 Sep 2023 01:43:36 +0000 en-US hourly 1 https://wordpress.org/?v=4.8.17 https://paper.nweon.com/wp-content/uploads/2021/04/nweon-icon.png Microsoft – Nweon Paper https://paper.nweon.com 32 32 VibHead: An Authentication Scheme for Smart Headsets through Vibration https://paper.nweon.com/14615 Thu, 27 Jul 2023 07:28:23 +0000 https://paper.nweon.com/14615 PubDate: June 2023

VibHead: An Authentication Scheme for Smart Headsets through Vibration最先出现在Nweon Paper

]]>
PubDate: June 2023

Teams: Shandong University;Qingdao University

Writers: Feng Li, Jiayi Zhao, Huan Yang, Dongxiao Yu, Yuanfeng Zhou, Yiran Shen

PDF: VibHead: An Authentication Scheme for Smart Headsets through Vibration

Abstract

Recent years have witnessed the fast penetration of Virtual Reality (VR) and Augmented Reality (AR) systems into our daily life, the security and privacy issues of the VR/AR applications have been attracting considerable attention. Most VR/AR systems adopt head-mounted devices (i.e., smart headsets) to interact with users and the devices usually store the users’ private data. Hence, authentication schemes are desired for the head-mounted devices. Traditional knowledge-based authentication schemes for general personal devices have been proved vulnerable to shoulder-surfing attacks, especially considering the headsets may block the sight of the users. Although the robustness of the knowledge-based authentication can be improved by designing complicated secret codes in virtual space, this approach induces a compromise of usability. Another choice is to leverage the users’ biometrics; however, it either relies on highly advanced equipments which may not always be available in commercial headsets or introduce heavy cognitive load to users.
In this paper, we propose a vibration-based authentication scheme, VibHead, for smart headsets. Since the propagation of vibration signals through human heads presents unique patterns for different individuals, VibHead employs a CNN-based model to classify registered legitimate users based the features extracted from the vibration signals. We also design a two-step authentication scheme where the above user classifiers are utilized to distinguish the legitimate user from illegitimate ones. We implement VibHead on a Microsoft HoloLens equipped with a linear motor and an IMU sensor which are commonly used in off-the-shelf personal smart devices. According to the results of our extensive experiments, with short vibration signals (≤1s), VibHead has an outstanding authentication accuracy; both FAR and FRR are around 5%.

VibHead: An Authentication Scheme for Smart Headsets through Vibration最先出现在Nweon Paper

]]>
ArK: Augmented Reality with Knowledge Interactive Emergent Ability https://paper.nweon.com/14363 Wed, 10 May 2023 04:51:15 +0000 https://paper.nweon.com/14363 PubDate: Apr 2023

ArK: Augmented Reality with Knowledge Interactive Emergent Ability最先出现在Nweon Paper

]]>
PubDate: Apr 2023

Teams: Microsoft Research, Redmond † MILA §University of Washington \ UCLA

Writers: Qiuyuan Huang, Jae Sung Park, Abhinav Gupta, Paul Bennett, Ran Gong, Subhojit Som, Baolin Peng, Owais Khan Mohammed, Chris Pal, Yejin Choi, Jianfeng Gao

PDF: ArK: Augmented Reality with Knowledge Interactive Emergent Ability

Abstract

Despite the growing adoption of mixed reality and interactive AI agents, it remains challenging for these systems to generate high quality 2D/3D scenes in unseen environments. The common practice requires deploying an AI agent to collect large amounts of data for model training for every new task. This process is costly, or even impossible, for many domains. In this study, we develop an infinite agent that learns to transfer knowledge memory from general foundation models (e.g. GPT4, DALLE) to novel domains or scenarios for scene understanding and generation in the physical or virtual world. The heart of our approach is an emerging mechanism, dubbed Augmented Reality with Knowledge Inference Interaction (ArK), which leverages knowledge-memory to generate scenes in unseen physical world and virtual reality environments. The knowledge interactive emergent ability (Figure 1) is demonstrated as the observation learns i) micro-action of cross-modality: in multi-modality models to collect a large amount of relevant knowledge memory data for each interaction task (e.g., unseen scene understanding) from the physical reality; and ii) macro-behavior of reality-agnostic: in mix-reality environments to improve interactions that tailor to different characterized roles, target variables, collaborative information, and so on. We validate the effectiveness of ArK on the scene generation and editing tasks. We show that our ArK approach, combined with large foundation models, significantly improves the quality of generated 2D/3D scenes, compared to baselines, demonstrating the potential benefit of incorporating ArK in generative AI for applications such as metaverse and gaming simulation.

ArK: Augmented Reality with Knowledge Interactive Emergent Ability最先出现在Nweon Paper

]]>
Inside-out Infrared Marker Tracking via Head Mounted Displays for Smart Robot Programming https://paper.nweon.com/14246 Thu, 06 Apr 2023 05:37:22 +0000 https://paper.nweon.com/14246 PubDate: Mar 2023

Inside-out Infrared Marker Tracking via Head Mounted Displays for Smart Robot Programming最先出现在Nweon Paper

]]>
PubDate: Mar 2023

Teams: Karlsruhe Institute of Technology;Karlsruhe University of Applied Sciences

Writers: David Puljiz, Alexandru-George Vasilache, Michael Mende, Björn Hein

PDF: Inside-out Infrared Marker Tracking via Head Mounted Displays for Smart Robot Programming

Abstract

Intuitive robot programming through use of tracked smart input devices relies on fixed, external tracking systems, most often employing infra-red markers. Such an approach is frequently combined with projector-based augmented reality for better visualisation and interface. The combined system, although providing an intuitive programming platform with short cycle times even for inexperienced users, is immobile, expensive and requires extensive calibration. When faced with a changing environment and large number of robots it becomes sorely impractical. Here we present our work on infra-red marker tracking using the Microsoft HoloLens head-mounted display. The HoloLens can map the environment, register the robot on-line, and track smart devices equipped with infra-red markers in the robot coordinate system. We envision our work to provide the basis to transfer many of the paradigms developed over the years for systems requiring a projector and a tracked input device into a highly-portable system that does not require any calibration or special set-up. We test the quality of the marker-tracking in an industrial robot cell and compare our tracking with a ground truth obtained via an ART-3 tracking system.

Inside-out Infrared Marker Tracking via Head Mounted Displays for Smart Robot Programming最先出现在Nweon Paper

]]>
Implementation of communication media around a mixed reality experience with HoloLens headset, as part of a digitalization of a nutrition workshop https://paper.nweon.com/14218 Wed, 29 Mar 2023 04:22:22 +0000 https://paper.nweon.com/14218 PubDate: Mar 2023

Implementation of communication media around a mixed reality experience with HoloLens headset, as part of a digitalization of a nutrition workshop最先出现在Nweon Paper

]]>
PubDate: Mar 2023

Teams: Université Clermont Auvergne;CNRS;Services de Chirurgie Bariatrique/Nutrition;Unité transversale d’éducation du patien;Unité de Recherche Clinique

Writers: Owen Kevin Appadoo, Hugo Rositi, Sylvie Valarier, Marie-Claire Ombret, Émilie Gadéa, Christine Barret-Grimault, Christophe Lohou

PDF: Implementation of communication media around a mixed reality experience with HoloLens headset, as part of a digitalization of a nutrition workshop

Abstract

The release of Microsoft’s HoloLens headset addresses new types of issues that would have been difficult to design without such a hardware. This semi-transparent visor headset allows the user who wears it to view the projection of 3D virtual objects placed in its real environment. The user can also interact with these 3D objects, which can interact with each other. The framework of this new technology is called mixed reality. We had the opportunity to numerically transform a conventional human nutrition workshop for patients waiting for bariatric surgery by developing a software called HOLO_NUTRI using the HoloLens headset. Unlike our experience of user and conventional programmer specialized in the development of interactive 3D graphics applications, we realized that such a mixed reality experience required specific programming concepts quite different from those of conventional software or those of virtual reality applications, but above all required a thorough reflection about communication for users. In this article, we propose to explain our design of communication (graphic supports, tutorials of use of material, explanatory videos), a step which was crucial for the good progress of our project. The software was used by thirty patients from Le Puy-en-Velay Hospital during 10 sessions of one hour and a half during which patients had to take in hand the headset and software HOLO_NUTRI. We also proposed a series of questions to patients to have an assessment of both the adequacy and the importance of this communication approach for such experience. As the mixed reality technology is very recent but the number of applications based on it significantly increases, the reflection on the implementation of the elements of communication described in this article (videos, exercise of learning for the use of the headset, communication leaflet, etc.) can help developers of such applications.

Implementation of communication media around a mixed reality experience with HoloLens headset, as part of a digitalization of a nutrition workshop最先出现在Nweon Paper

]]>
Augmenting Augmented Reality with Non-Line-of-Sight Perception https://paper.nweon.com/14114 Tue, 28 Feb 2023 01:13:39 +0000 https://paper.nweon.com/14114 PubDate: Feb 2023

Augmenting Augmented Reality with Non-Line-of-Sight Perception最先出现在Nweon Paper

]]>
PubDate: Feb 2023

Teams: 1 Massachusetts Institute of Technology, 2 University of Michigan

Writers: Tara Boroushaki 1, Maisy Lam 1, Laura Dodds 1, Aline Eid 1,2, Fadel Adib 1

PDF: Augmenting Augmented Reality with Non-Line-of-Sight Perception

Abstract

We present the design, implementation, and evaluation of X-AR, an augmented reality (AR) system with non-line-of-sight perception. X-AR augments AR headsets with RF sensing to enable users to see things that are otherwise invisible to the human eye or to stateof-the-art AR systems. Our design introduces three main innovations: the first is an AR-conformal antenna that tightly matches the shape of the AR headset visor while providing excellent radiation and bandwidth capabilities for RF sensing. The second is an RF-visual synthetic aperture localization algorithm that leverages natural human mobility to localize RF-tagged objects in line-ofsight and non-line-of-sight settings. Finally, the third is an RF-visual verification primitive that fuses RF and vision to deliver actionable tasks to end users such as picking verification. We built an end-to-end prototype of our design by integrating it into a Microsoft Hololens 2 AR headset and evaluated it in line-of-sight and nonline-of-sight environments. Our results demonstrate that X-AR achieves decimeter-level RF localization (median of 9.8 cm) of fully-occluded items and can perform RFvisual picking verification with over 95% accuracy (FScore) when extracting RFID-tagged items.These results show that X-AR is successful in extending AR systems to non-line-of-sight perception, with important implications to manufacturing, warehousing, and smart home applications. Demo video: y2u.be/bdUN21ft7G0

Augmenting Augmented Reality with Non-Line-of-Sight Perception最先出现在Nweon Paper

]]>
Spatial audio signal processing for binaural reproduction of recorded acoustic scenes – review and challenges https://paper.nweon.com/14016 Tue, 31 Jan 2023 05:46:28 +0000 https://paper.nweon.com/14016 PubDate: Oct 2022

Spatial audio signal processing for binaural reproduction of recorded acoustic scenes – review and challenges最先出现在Nweon Paper

]]>
PubDate: Oct 2022

Teams: Ben-Gurion University of the Negev,Meta;International Audio Laboratories Erlangen;University of Huddersfield;, Microsoft Research;The Australian National University

Writers: Boaz Rafaely, Vladimir Tourbabin, Emanuel Habets, Zamir Ben-Hur, Hyunkook Lee, Hannes Gamper, Lior Arbel, Lachlan Birnie, Thushara Abhayapala, Prasanga Samarasinghe

PDF: Spatial audio signal processing for binaural reproduction of recorded acoustic scenes – review and challenges

Abstract

Spatial audio has been studied for several decades, but has seen much renewed interest recently due to advances in both software and hardware for capture and playback, and the emergence of applications such as virtual reality and augmented reality. This renewed interest has led to the investment of increasing efforts in developing signal processing algorithms for spatial audio, both for capture and for playback. In particular, due to the popularity of headphones and earphones, many spatial audio signal processing methods have dealt with binaural reproduction based on headphone listening. Among these new developments, processing spatial audio signals recorded in real environments using microphone arrays plays an important role. Following this emerging activity, this paper aims to provide a scientific review of recent developments and an outlook for future challenges. This review also proposes a generalized framework for describing spatial audio signal processing for the binaural reproduction of recorded sound. This framework helps to understand the collective progress of the research community, and to identify gaps for future research. It is composed of five main blocks, namely: the acoustic scene, recording, processing, reproduction, and perception and evaluation. First, each block is briefly presented, and then, a comprehensive review of the processing block is provided. This includes topics from simple binaural recording to Ambisonics and perceptually motivated approaches, which focus on careful array configuration and design. Beamforming and parametric-based processing afford more flexible designs and shift the focus to processing and modeling of the sound field. Then, emerging machine- and deep-learning approaches, which take a further step towards flexibility in design, are described. Finally, specific methods for signal transformations such as rotation, translation and enhancement, enabling additional flexibility in reproduction and improvement in the quality of the binaural signal, are presented. The review concludes by highlighting directions for future research.

Spatial audio signal processing for binaural reproduction of recorded acoustic scenes – review and challenges最先出现在Nweon Paper

]]>
Imitator: Personalized Speech-driven 3D Facial Animation https://paper.nweon.com/13838 Wed, 18 Jan 2023 04:31:20 +0000 https://paper.nweon.com/13838 PubDate: Dec 2022

Imitator: Personalized Speech-driven 3D Facial Animation最先出现在Nweon Paper

]]>
PubDate: Dec 2022

Teams: Max Planck Institute for Intelligent Systems;Microsoft Mixed Reality & AI Lab

Writers: Balamurugan Thambiraja, Ikhsanul Habibie, Sadegh Aliakbarian, Darren Cosker, Christian Theobalt, Justus Thies

PDF: Imitator: Personalized Speech-driven 3D Facial Animation

Abstract

Speech-driven 3D facial animation has been widely explored, with applications in gaming, character animation, virtual reality, and telepresence systems. State-of-the-art methods deform the face topology of the target actor to sync the input audio without considering the identity-specific speaking style and facial idiosyncrasies of the target actor, thus, resulting in unrealistic and inaccurate lip movements. To address this, we present Imitator, a speech-driven facial expression synthesis method, which learns identity-specific details from a short input video and produces novel facial expressions matching the identity-specific speaking style and facial idiosyncrasies of the target actor. Specifically, we train a style-agnostic transformer on a large facial expression dataset which we use as a prior for audio-driven facial expressions. Based on this prior, we optimize for identity-specific speaking style based on a short reference video. To train the prior, we introduce a novel loss function based on detected bilabial consonants to ensure plausible lip closures and consequently improve the realism of the generated expressions. Through detailed experiments and a user study, we show that our approach produces temporally coherent facial expressions from input audio while preserving the speaking style of the target actors.

Imitator: Personalized Speech-driven 3D Facial Animation最先出现在Nweon Paper

]]>
Gaze-Vergence-Controlled See-Through Vision in Augmented Reality https://paper.nweon.com/13647 Wed, 07 Dec 2022 02:16:45 +0000 https://paper.nweon.com/?p=13647 ...

Gaze-Vergence-Controlled See-Through Vision in Augmented Reality最先出现在Nweon Paper

]]>
PubDate: Sep 2022

Teams:  Beihang University

Writers: Zhimin Wang; Yuxin Zhao; Feng Lu

PDFGaze-Vergence-Controlled See-Through Vision in Augmented Reality

Abstract

Augmented Reality (AR) see-through vision is an interesting research topic since it enables users to see through a wall and see the occluded objects. Most existing research focuses on the visual effects of see-through vision, while the interaction method is less studied. However, we argue that using common interaction modalities, e.g., midair click and speech, may not be the optimal way to control see-through vision. This is because when we want to see through something, it is physically related to our gaze depth/vergence and thus should be naturally controlled by the eyes. Following this idea, this paper proposes a novel gaze-vergence-controlled (GVC) see-through vision technique in AR. Since gaze depth is needed, we build a gaze tracking module with two infrared cameras and the corresponding algorithm and assemble it into the Microsoft HoloLens 2 to achieve gaze depth estimation. We then propose two different GVC modes for see-through vision to fit different scenarios. Extensive experimental results demonstrate that our gaze depth estimation is efficient and accurate. By comparing with conventional interaction modalities, our GVC techniques are also shown to be superior in terms of efficiency and more preferred by users. Finally, we present four example applications of gaze-vergence-controlled see-through vision.

Gaze-Vergence-Controlled See-Through Vision in Augmented Reality最先出现在Nweon Paper

]]>
LaMAR: Benchmarking Localization and Mapping for Augmented Reality https://paper.nweon.com/13300 Mon, 24 Oct 2022 13:37:25 +0000 https://paper.nweon.com/13300 PubDate: Oct 2022

LaMAR: Benchmarking Localization and Mapping for Augmented Reality最先出现在Nweon Paper

]]>
PubDate: Oct 2022

Teams: ETH Zurich;Microsoft Mixed Reality & AI Lab

Writers: Paul-Edouard Sarlin, Mihai-Alexandru Dusmanu, Johannes L. Schönberger, Pablo Speciale, Lukas Gruber, Viktor Larsson, Ondrej Miksik, Marc Pollefeys

PDF: LaMAR: Benchmarking Localization and Mapping for Augmented Reality

Abstract

Localization and mapping is the foundational technology for augmented reality (AR) that enables sharing and persistence of digital content in the real world. While significant progress has been made, researchers are still mostly driven by unrealistic benchmarks not representative of real-world AR scenarios. In particular, benchmarks are often based on small-scale datasets with low scene diversity, captured from stationary cameras, and lacking other sensor inputs like inertial, radio, or depth data. Furthermore, ground-truth (GT) accuracy is mostly insufficient to satisfy AR requirements. To close this gap, we introduce a new benchmark with a comprehensive capture and GT pipeline, which allow us to co-register realistic AR trajectories in diverse scenes and from heterogeneous devices at scale. To establish accurate GT, our pipeline robustly aligns the captured trajectories against laser scans in a fully automatic manner. Based on this pipeline, we publish a benchmark dataset of diverse and large-scale scenes recorded with head-mounted and hand-held AR devices. We extend several state-of-the-art methods to take advantage of the AR specific setup and evaluate them on our benchmark. Based on the results, we present novel insights on current research gaps to provide avenues for future work in the community.

LaMAR: Benchmarking Localization and Mapping for Augmented Reality最先出现在Nweon Paper

]]>
3D Face Reconstruction with Dense Landmarks https://paper.nweon.com/13298 Mon, 24 Oct 2022 13:37:23 +0000 https://paper.nweon.com/13298 PubDate: Oct 2022

3D Face Reconstruction with Dense Landmarks最先出现在Nweon Paper

]]>
PubDate: Oct 2022

Teams: Microsoft

Writers: Erroll Wood Tadas Baltrušaitis Charlie Hewitt Matthew Johnson Jingjing Shen Nikola Milosavljevic Daniel Wilde Stephan Garbin Chirag Raman Jamie Shotton Toby Sharp Ivan Stojiljkovic Thomas J. Cashman Julien Valentin

PDF: 3D Face Reconstruction with Dense Landmarks

Abstract

Landmarks often play a key role in face analysis, but many aspects of identity or expression cannot be represented by sparse landmarks alone. Thus, in order to reconstruct faces more accurately, landmarks are often combined with additional signals like depth images or techniques like differentiable rendering.

Can we keep things simple by just using more landmarks?

In answer, we present the first method that accurately predicts ten times as many landmarks as usual, covering the whole head, including the eyes and teeth. This is accomplished using synthetic training data, which guarantees perfect landmark annotations. By fitting a morphable model to these dense landmarks, we achieve state-of-the-art results for monocular 3D face reconstruction in the wild. We show that dense landmarks are an ideal signal for integrating face shape information across frames by demonstrating accurate and expressive facial performance capture in both monocular and multi-view scenarios. Finally, our method is highly efficient: we can predict dense landmarks and fit our 3D face model at over 150FPS on a single CPU thread.

3D Face Reconstruction with Dense Landmarks最先出现在Nweon Paper

]]>
Attention Guidance for Tower ATC Using Augmented Reality Devices https://paper.nweon.com/13157 Wed, 21 Sep 2022 04:13:22 +0000 https://paper.nweon.com/13157 PubDate: May 2022

Attention Guidance for Tower ATC Using Augmented Reality Devices最先出现在Nweon Paper

]]>
PubDate: May 2022

Teams: Royal Netherlands Aerospace Centre

Writers: Jürgen Teutsch; Tanja Bos; Marcel van Apeldoorn; Lansenou Camara

PDF: Attention Guidance for Tower ATC Using Augmented Reality Devices

Abstract

In 2021 Royal NLR carried out innovative technology experiments on their high-fidelity real-time air traffic control simulation and validation platform, NARSIM. These experiments were part of the SESAR 2020 project Digital Technologies for Tower (DTT). The technology option that was investigated focused on advanced HMI interaction modes for aerodrome tower controllers. More particular, Attention Capturing and Guidance strategies with an Augmented Reality device, the Microsoft HoloLens 2™, were evaluated inside an aerodrome control tower environment for Amsterdam Airport Schiphol, one of the major European hub airports.The NARSIM environment consisted of a realistic but downscaled presentation of the airport with two tower controller working positions emulating current tower systems. Such a set-up allowed researchers to focus their work on the application of Augmented Reality with the introduction of (virtual) aircraft labels as well as special symbology and auditory cues for capturing and guiding tower controller attention in the case of critical events. Several typical attention-critical events that may occur at an airport, such as go-around operations and runway incursions, were orchestrated by a team of NLR experts and presented to the tower controllers while they were operating traffic as usual. Human performance and ATC operational experts observed and analyzed the simulations.This paper describes the steps taken and the challenges encountered when integrating the HoloLens inside the NARSIM Tower environment. Furthermore, it explores the proposed operational concept for Attention Capturing and Guidance with the HoloLens and how it was realized inside the device. The results of the technical evaluation activity with two experienced air traffic controllers are described in detail. These results came to the conclusion that the device in combination with the concept was a favorable addition to the controller working environment. While desired technical performance improvements, mostly related to user comfort and general adjustments, depend on further vendor development, the used HoloLens was seen as a technically useful device for implementing prototype applications for Attention Capturing and Guidance with aural and visual cues.In the final sections of the paper, an outlook into the expected future use of Augmented Reality devices in conventional control tower environments is given.

Attention Guidance for Tower ATC Using Augmented Reality Devices最先出现在Nweon Paper

]]>
Control with Vergence Eye Movement in Augmented Reality See-Through Vision https://paper.nweon.com/13024 Tue, 06 Sep 2022 02:19:21 +0000 https://paper.nweon.com/13024 PubDate: April 2022

Control with Vergence Eye Movement in Augmented Reality See-Through Vision最先出现在Nweon Paper

]]>
PubDate: April 2022

Teams: Beihang University

Writers: Zhimin Wang; Yuxin Zhao; Feng Lu

PDF: Control with Vergence Eye Movement in Augmented Reality See-Through Vision

Abstract

Augmented Reality (AR) see-through vision has become a recent research focus since it enables the user to see through a wall and see the occluded objects. Most existing works only used common modalities to control the display for see-through vision, e.g., button clicking and speech control. However, we use visual system to observe see-through vision. Using an addition interaction channel will distract the user and degrade the user experience. In this paper, we propose a novel interaction method using vergence eye movement for controlling see-through vision in AR. Specifically, we first customize eye cameras and design gaze depth estimation method for Microsoft HoloLens 2. With our algorithm, fixation depth can be computed from the vergence, and used to manage the see-through vision. We also propose two control techniques of gaze vergence. The experimental results show that the gaze depth estimation method is efficient. The difference cannot be found between these two modalities in terms of completion time and the number of successes.

Control with Vergence Eye Movement in Augmented Reality See-Through Vision最先出现在Nweon Paper

]]>
A Pilot Study on The Impact of Stereoscopic Display Type on User Interactions Within A Immersive Analytics Environment https://paper.nweon.com/12940 Wed, 31 Aug 2022 04:31:19 +0000 https://paper.nweon.com/12940 PubDate: Jul 2022

A Pilot Study on The Impact of Stereoscopic Display Type on User Interactions Within A Immersive Analytics Environment最先出现在Nweon Paper

]]>
PubDate: Jul 2022

Teams: Colorado State University;Microsoft Research

Writers: Adam S. Williams, Xiaoyan Zhou, Michel Pahud, Francisco R. Ortega

PDF: A Pilot Study on The Impact of Stereoscopic Display Type on User Interactions Within A Immersive Analytics Environment

Abstract

Immersive Analytics (IA) and consumer adoption of augmented reality (AR) and virtual reality (VR) head-mounted displays (HMDs) are both rapidly growing. When used in conjunction, stereoscopic IA environments can offer improved user understanding and engagement; however, it is unclear how the choice of stereoscopic display impacts user interactions within an IA environment. This paper presents a pilot study that examines the impact of stereoscopic display type on object manipulation and environmental navigation using consumer-available AR and VR displays. This work finds that the display type can impact how users manipulate virtual content, how they navigate the environment, and how able they are to answer questions about the represented data.

A Pilot Study on The Impact of Stereoscopic Display Type on User Interactions Within A Immersive Analytics Environment最先出现在Nweon Paper

]]>
Path Tracing in 2D, 3D, and Physicalized Networks https://paper.nweon.com/12934 Wed, 31 Aug 2022 02:34:20 +0000 https://paper.nweon.com/12934 PubDate: Jul 2022

Path Tracing in 2D, 3D, and Physicalized Networks最先出现在Nweon Paper

]]>
PubDate: Jul 2022

Teams: Ecole de technologie sup´erieure

Writers: Michael J. McGuffin (ETS), Ryan Servera (ETS), Marie Forest (ETS)

PDF: Path Tracing in 2D, 3D, and Physicalized Networks

Abstract

It is common to advise against using 3D to visualize abstract data such as networks, however Ware and Mitchell’s 2008 study showed that path tracing in a network is less error prone in 3D than in 2D. It is unclear, however, if 3D retains its advantage when the 2D presentation of a network is improved using edge-routing, and when simple interaction techniques for exploring the network are available. We address this with two studies of path tracing under new conditions. The first study was preregistered, involved 34 users, and compared 2D and 3D layouts that the user could rotate and move in virtual reality with a handheld controller. Error rates were lower in 3D than in 2D, despite the use of edge-routing in 2D and the use of mouse-driven interactive highlighting of edges. The second study involved 12 users and investigated data physicalization, comparing 3D layouts in virtual reality versus physical 3D printouts of networks augmented with a Microsoft HoloLens headset. No difference was found in error rate, but users performed a variety of actions with their fingers in the physical condition which can inform new interaction techniques.

Path Tracing in 2D, 3D, and Physicalized Networks最先出现在Nweon Paper

]]>
MoCapAct: A Multi-Task Dataset for Simulated Humanoid Control https://paper.nweon.com/12909 Sun, 28 Aug 2022 23:43:48 +0000 https://paper.nweon.com/12909 PubDate: Aug 2022

MoCapAct: A Multi-Task Dataset for Simulated Humanoid Control最先出现在Nweon Paper

]]>
PubDate: Aug 2022

Teams: Georgia Institute of Technology;Microsoft Research

Writers: Nolan Wagener, Andrey Kolobov, Felipe Vieira Frujeri, Ricky Loynd, Ching-An Cheng, Matthew Hausknecht

PDF: MoCapAct: A Multi-Task Dataset for Simulated Humanoid Control

Abstract

Simulated humanoids are an appealing research domain due to their physical capabilities. Nonetheless, they are also challenging to control, as a policy must drive an unstable, discontinuous, and high-dimensional physical system. One widely studied approach is to utilize motion capture (MoCap) data to teach the humanoid agent low-level skills (e.g., standing, walking, and running) that can then be re-used to synthesize high-level behaviors. However, even with MoCap data, controlling simulated humanoids remains very hard, as MoCap data offers only kinematic information. Finding physical control inputs to realize the demonstrated motions requires computationally intensive methods like reinforcement learning. Thus, despite the publicly available MoCap data, its utility has been limited to institutions with large-scale compute. In this work, we dramatically lower the barrier for productive research on this topic by training and releasing high-quality agents that can track over three hours of MoCap data for a simulated humanoid in the dm_control physics-based environment. We release MoCapAct (Motion Capture with Actions), a dataset of these expert agents and their rollouts, which contain proprioceptive observations and actions. We demonstrate the utility of MoCapAct by using it to train a single hierarchical policy capable of tracking the entire MoCap dataset within dm_control and show the learned low-level component can be re-used to efficiently learn downstream high-level tasks. Finally, we use MoCapAct to train an autoregressive GPT model and show that it can control a simulated humanoid to perform natural motion completion given a motion prompt.
Videos of the results and links to the code and dataset are available at this https URL.

MoCapAct: A Multi-Task Dataset for Simulated Humanoid Control最先出现在Nweon Paper

]]>
Depth Perception in Augmented Reality: The Effects of Display, Shadow, and Position https://paper.nweon.com/12866 Wed, 24 Aug 2022 06:16:23 +0000 https://paper.nweon.com/12866 PubDate: April 2022

Depth Perception in Augmented Reality: The Effects of Display, Shadow, and Position最先出现在Nweon Paper

]]>
PubDate: April 2022

Teams: Vanderbilt University;University of Utah

Writers: Haley Adams; Jeanine Stefanucci; Sarah Creem-Regehr; Bobby Bodenheimer

PDF: Depth Perception in Augmented Reality: The Effects of Display, Shadow, and Position

Abstract

Although it is commonly accepted that depth perception in augmented reality (AR) displays is distorted, we have yet to isolate which properties of AR affect people’s ability to correctly perceive virtual objects in real spaces. From prior research on depth perception in commercial virtual reality, it is likely that ergonomic properties and graphical limitations impact visual perception in head-mounted displays (HMDs). However, an insufficient amount of research has been conducted in augmented reality HMDs for us to begin isolating pertinent factors in this family of displays. To this end, in the current research, we evaluate absolute measures of distance perception in the Microsoft HoloLens 2, an optical see-through AR display, and the Varjo XR-3, a video see-through AR display. The current work is the first to evaluate either device using absolute distance perception as a measure. For each display, we asked participants to verbally report distance judgments to both grounded and floating targets that were rendered either with or without a cast shadow along the ground. Our findings suggest that currently available video see-through displays may induce more distance underestimation than their optical see-through counterparts. We also find that the vertical position of an object and the presence of a cast shadow influence depth perception.

Depth Perception in Augmented Reality: The Effects of Display, Shadow, and Position最先出现在Nweon Paper

]]>
Contactless Haptic Display Through Magnetic Field Control https://paper.nweon.com/12773 Tue, 16 Aug 2022 07:55:38 +0000 https://paper.nweon.com/12773 PubDate: February 2022

Contactless Haptic Display Through Magnetic Field Control最先出现在Nweon Paper

]]>
PubDate: February 2022

Teams: Nanjing University of Aeronautics and Astronautics;UNSW

Writers: Xiong Lu; Yuxing Yan; Beibei Qi; Huang Qian; Junbin Sun; Aaron Quigley

PDF: Contactless Haptic Display Through Magnetic Field Control

Abstract

Haptic rendering enables people to touch, perceive, and manipulate virtual objects in a virtual environment. Using six cascaded identical hollow disk electromagnets and a small permanent magnet attached to an operator’s finger, this paper proposes and develops an untethered haptic interface through magnetic field control. The concentric hole inside the six cascaded electromagnets provides the workspace, where the 3D position of the permanent magnet is tracked with a Microsoft Kinect sensor. The driving currents of six cascaded electromagnets are calculated in real-time for generating the desired magnetic force. Offline data from an FEA (finite element analysis) based simulation, determines the relationship between the magnetic force, the driving currents, and the position of the permanent magnet. A set of experiments including the virtual object recognition experiment, the virtual surface identification experiment, and the user perception evaluation experiment were conducted to demonstrate the proposed system, where Microsoft HoloLens holographic glasses are used for visual rendering. The proposed magnetic haptic display leads to an untethered and non-contact interface for natural haptic rendering applications, which overcomes the constraints of mechanical linkages in tool-based traditional haptic devices.

Contactless Haptic Display Through Magnetic Field Control最先出现在Nweon Paper

]]>
Measuring understorey vegetation structure using a novel mixed-reality device https://paper.nweon.com/12613 Wed, 13 Jul 2022 23:46:21 +0000 https://paper.nweon.com/12613 PubDate: June 2022

Measuring understorey vegetation structure using a novel mixed-reality device最先出现在Nweon Paper

]]>
PubDate: June 2022

Teams: Rice University

Writers: Daniel Gorczynski,Lydia Beaudrot

PDF: Measuring understorey vegetation structure using a novel mixed-reality device

Abstract

Most ecological studies of vegetation structure have relied on manual field measurements that are labour-intensive and time-consuming. Many current alternatives to classical measurements are expensive or difficult to transport to field settings.
Here we evaluated a new method for measuring understorey vegetation with a novel mixed-reality, remote sensing device, the Microsoft HoloLens. We developed a vegetation sensing application called VegSense that allows the HoloLens user to control the device’s environmental scanners to measure understorey vegetation. Using VegSense, we tested the ability of the Microsoft HoloLens relative to classical field measurements to (a) detect trees and saplings, (b) measure diameter at breast height (DBH), (c) detect individual understorey vegetation structures and (d) estimate understorey vegetation complexity replicating the rod-transect method.
We found that VegSense performed well in detecting and measuring trees with a DBH of 17 cm or more and estimating vegetation complexity and performed moderately at detecting understorey vegetation.
Our results indicate that the HoloLens is a suitable alternative for multiple classical field measurements of understorey vegetation. This method costs much less than typical terrestrial LiDAR systems, and can facilitate efficient, high-quality environmental data collection. Further software development has the potential to reveal additional ways in which this device can be harnessed for applications to ecology and evolution.

Measuring understorey vegetation structure using a novel mixed-reality device最先出现在Nweon Paper

]]>
A Fast Forest Reverberator Using Single Scattering Cylinders https://paper.nweon.com/12573 Tue, 12 Jul 2022 05:52:19 +0000 https://paper.nweon.com/12573 PubDate: March 2022

A Fast Forest Reverberator Using Single Scattering Cylinders最先出现在Nweon Paper

]]>
PubDate: March 2022

Teams: University of Maryland;Microsoft Research

Writers: Shoken Kaneko; Hannes Gamper

PDF: A Fast Forest Reverberator Using Single Scattering Cylinders

Abstract

Simulating forest acoustics has important applications for rendering forest sound scenes in mixed and virtual reality, developing wildlife monitoring systems that use microphone arrays distributed in a forest, or as an artistic effect. Previously proposed methods for forest impulse response (IR) synthesis are limited to small or sparse forests because of their cubic asymptotic complexity with respect to the number of trees. Here we propose a simple and efficient parametric forest IR generation algorithm that relies on a multitude of single scattering cylinders to approximate scattering caused by tree trunks. The proposed method was compared to measured forest IRs in terms of the IR echo density, energy decay, reverberation time (T60), and clarity (C50). Experimental results indicate that the proposed algorithm generates forest reverb with acoustic characteristics similar to real forest IRs at a low computational cost.

A Fast Forest Reverberator Using Single Scattering Cylinders最先出现在Nweon Paper

]]>
A 3D Reconstruction Method for Augmented Reality Sandbox Based on Depth Sensor https://paper.nweon.com/12534 Mon, 11 Jul 2022 00:58:29 +0000 https://paper.nweon.com/12534 PubDate: February 2022

A 3D Reconstruction Method for Augmented Reality Sandbox Based on Depth Sensor最先出现在Nweon Paper

]]>
PubDate: February 2022

Teams: Shanghai Branch of National Computer network Emergency Response technical Team

Writers: Xindan Wang; Qu Chen; Zhi Li

PDF: A 3D Reconstruction Method for Augmented Reality Sandbox Based on Depth Sensor

Abstract

This paper builds an Augmented Reality Sandbox (AR Sandbox) system based on augmented reality technology, and performs a 3D reconstruction for the sandbox terrain using the depth sensor Microsoft Kinect in the AR Sandbox, as an entry point to pave the way for later development of related metaverse applications, such as the metaverse architecting and visual interactive modeling. The innovation of this paper is that for the AR Sandbox scene, a 3D reconstruction method based on depth sensor is proposed, which can automatically cut off the edge of the sandbox table in Kinect field of view, and accurately and completely reconstruct the sandbox terrain in Matlab.

A 3D Reconstruction Method for Augmented Reality Sandbox Based on Depth Sensor最先出现在Nweon Paper

]]>
Synthesis of User-Level 3D Head Avatar from Smartphone Video for Virtual Reality https://paper.nweon.com/12503 Wed, 06 Jul 2022 07:25:20 +0000 https://paper.nweon.com/12503 PubDate: January 2022

Synthesis of User-Level 3D Head Avatar from Smartphone Video for Virtual Reality最先出现在Nweon Paper

]]>
PubDate: January 2022

Teams: Bangladesh University of Engineering and Technology

Writers: Md Mushfiqur Rahman; S. M. Mahbubur Rahman

PDF: Synthesis of User-Level 3D Head Avatar from Smartphone Video for Virtual Reality

Abstract

The purpose of an avatar is to intensify the sense of reality in a simulated environment such as a virtual reality system. This is done by creating a replica of a real person in virtual reality, as intricately as possible. Implementation of 3D scanners (i.e., Microsoft Kinect) produced vast opportunities for creating 3D models of celebrities and other people who can afford them. These humanoid 3D models as well as fictional human avatars opened the door of perceiving a more realistic 3D system than before. However, a simplified approach of fully rigged 3D head avatar synthesis may help to expand the application of 3D avatars allowing anyone to feel a personalized experience of their own avatar. In this work, we propose a system of 3D head avatar synthesis at the user-end, using an easily accessible and low-cost sensor (smartphone camera). Our proposed methodology consists of data acquisition of the concerned person as video, feature extraction and matching, image registration, formation of structure form motion and post-processing. Without requiring any training data, our method provides a satisfactory performance based on the experimental results. Additionally, we release a public database for avatar synthesis from smartphone video and introduce a set of attributes to evaluate the synthesized 3D avatars.

Synthesis of User-Level 3D Head Avatar from Smartphone Video for Virtual Reality最先出现在Nweon Paper

]]>
Quantifying the Effects of Working in VR for One Week https://paper.nweon.com/12389 Thu, 16 Jun 2022 02:16:25 +0000 https://paper.nweon.com/12389 PubDate: Jun 2022

Quantifying the Effects of Working in VR for One Week最先出现在Nweon Paper

]]>
PubDate: Jun 2022

Teams: Coburg University of Applied Sciences;Microsoft Research;University of Cambridge;University of Primorska

Writers: Verena Biener, Snehanjali Kalamkar, Negar Nouri, Eyal Ofek, Michel Pahud, John J. Dudley, Jinghui Hu, Per Ola Kristensson, Maheshya Weerasinghe, Klen Čopič Pucihar, Matjaž Kljun, Stephan Streuber, Jens Grubert

PDF: Quantifying the Effects of Working in VR for One Week

Abstract

Virtual Reality (VR) provides new possibilities for modern knowledge work. However, the potential advantages of virtual work environments can only be used if it is feasible to work in them for an extended period of time. Until now, there are limited studies of long-term effects when working in VR. This paper addresses the need for understanding such long-term effects. Specifically, we report on a comparative study (n=16), in which participants were working in VR for an entire week – for five days, eight hours each day – as well as in a baseline physical desktop environment. This study aims to quantify the effects of exchanging a desktop-based work environment with a VR-based environment. Hence, during this study, we do not present the participants with the best possible VR system but rather a setup delivering a comparable experience to working in the physical desktop environment. The study reveals that, as expected, VR results in significantly worse ratings across most measures. Among other results, we found concerning levels of simulator sickness, below average usability ratings and two participants dropped out on the first day using VR, due to migraine, nausea and anxiety. Nevertheless, there is some indication that participants gradually overcame negative first impressions and initial discomfort. Overall, this study helps lay the groundwork for subsequent research, by clearly highlighting current shortcomings and identifying opportunities for improving the experience of working in VR.

Quantifying the Effects of Working in VR for One Week最先出现在Nweon Paper

]]>
BinauralGrad: A Two-Stage Conditional Diffusion Probabilistic Model for Binaural Audio Synthesis https://paper.nweon.com/12345 Mon, 06 Jun 2022 07:52:26 +0000 https://paper.nweon.com/12345 PubDate: May 2022

BinauralGrad: A Two-Stage Conditional Diffusion Probabilistic Model for Binaural Audio Synthesis最先出现在Nweon Paper

]]>
PubDate: May 2022

Teams: University of Science and Technology of China;Microsoft Research Asia;Imperial College London;Microsoft Azure Speech;University of Surrey;South China University of Technology

Writers: Yichong Leng, Zehua Chen, Junliang Guo, Haohe Liu, Jiawei Chen, Xu Tan, Danilo Mandic, Lei He, Xiang-Yang Li, Tao Qin, Sheng Zhao, Tie-Yan Liu

PDF: BinauralGrad: A Two-Stage Conditional Diffusion Probabilistic Model for Binaural Audio Synthesis

Abstract

Binaural audio plays a significant role in constructing immersive augmented and virtual realities. As it is expensive to record binaural audio from the real world, synthesizing them from mono audio has attracted increasing attention. This synthesis process involves not only the basic physical warping of the mono audio, but also room reverberations and head/ear related filtrations, which, however, are difficult to accurately simulate in traditional digital signal processing. In this paper, we formulate the synthesis process from a different perspective by decomposing the binaural audio into a common part that shared by the left and right channels as well as a specific part that differs in each channel. Accordingly, we propose BinauralGrad, a novel two-stage framework equipped with diffusion models to synthesize them respectively. Specifically, in the first stage, the common information of the binaural audio is generated with a single-channel diffusion model conditioned on the mono audio, based on which the binaural audio is generated by a two-channel diffusion model in the second stage. Combining this novel perspective of two-stage synthesis with advanced generative models (i.e., the diffusion models),the proposed BinauralGrad is able to generate accurate and high-fidelity binaural audio samples. Experiment results show that on a benchmark dataset, BinauralGrad outperforms the existing baselines by a large margin in terms of both object and subject evaluation metrics (Wave L2: 0.128 vs. 0.157, MOS: 3.80 vs. 3.61). The generated audio samples are available online.

BinauralGrad: A Two-Stage Conditional Diffusion Probabilistic Model for Binaural Audio Synthesis最先出现在Nweon Paper

]]>
DreamStream: Immersive and Interactive Spectating in VR https://paper.nweon.com/12294 Wed, 25 May 2022 07:13:25 +0000 https://paper.nweon.com/12294 PubDate:

DreamStream: Immersive and Interactive Spectating in VR最先出现在Nweon Paper

]]>
PubDate:

Teams: Microsoft Research

Writers: Balasaravanan Thoravi Kumaravel;Andrew D Wilson

PDF: DreamStream: Immersive and Interactive Spectating in VR

Abstract

Today spectating and streaming virtual reality (VR) activities typically involves spectators viewing a 2D stream of the VR user’s view. Streaming 2D videos of the game play is popular and well-supported by platforms such as Twitch. However, the generic streaming of full 3D representations is less explored. Thus, while the VR player’s experience may be fully immersive, spectators are limited to 2D videos. This asymmetry lessens the overall experience for spectators, who themselves may be eager to spectate in VR. DreamStream puts viewers in the virtual environment of the VR application, allowing them to look “over the shoulder” of the VR player. Spectators can view streamed VR content immersively in 3D, independently explore the VR scene beyond what the VR player sees and ultimately cohabit the virtual environment alongside the VR player. For the VR player, DreamStream provides a spatial awareness of all their spectators. DreamStream retrofits and works with existing VR applications. We discuss the design and implementation of DreamStream, and carry out three qualitative informal evaluations. These evaluations shed light on the strengths and weakness of using DreamStream for the purpose of interactive spectating. Our participants found that DreamStream’s VR viewer interface offered increased immersion, and made it easier to communicate and interact with the VR player.

DreamStream: Immersive and Interactive Spectating in VR最先出现在Nweon Paper

]]>
Rotation-constrained optical see-through headset calibration with bare-hand alignment https://paper.nweon.com/12062 Thu, 21 Apr 2022 06:13:21 +0000 https://paper.nweon.com/12062 PubDate:

Rotation-constrained optical see-through headset calibration with bare-hand alignment最先出现在Nweon Paper

]]>
PubDate:

Teams: Imperial College London

Writers: Xue Hu; Ferdinando Rodriguez y Baena; Fabrizio Cutolo

PDF: Rotation-constrained optical see-through headset calibration with bare-hand alignment

Abstract

The inaccessibility of user-perceived reality remains an open issue in pursuing the accurate calibration of optical see-through (OST) head-mounted displays (HMDs). Manual user alignment is usually required to collect a set of virtual-to-real correspondences, so that a default or an offline display calibration can be updated to account for the user’s eye position(s). Current alignment-based calibration procedures usually require point-wise alignments between rendered image point(s) and associated physical landmark(s) of a target calibration tool. As each alignment can only provide one or a few correspondences, repeated alignments are required to ensure calibration quality. This work presents an accurate and tool-less online OST calibration method to update an offline-calibrated eye-display model. The user’s bare hand is markerlessly tracked by a commercial RGBD camera anchored to the OST headset to generate a user-specific cursor for correspondence collection. The required alignment is object-wise, and can provide thousands of unordered corresponding points in tracked space. The collected correspondences are registered by a proposed rotation-constrained iterative closest point (rcICP) method to optimise the viewpoint-related calibration parameters. We implemented such a method for the Microsoft HoloLens 1. The resiliency of the proposed procedure to noisy data was evaluated through simulated tests and real experiments performed with an eye-replacement camera. According to the simulation test, the rcICP registration is robust against possible user-induced rotational misalignment. With a single alignment, our method achieves 8.81 arcmin (1.37 mm) positional error and 1. 76° rotational error by camera-based tests in the arm-reach distance, and 10.79 arcmin (7.71 pixels) reprojection error by user tests.

Rotation-constrained optical see-through headset calibration with bare-hand alignment最先出现在Nweon Paper

]]>
Measuring the Perceived Three-Dimensional Location of Virtual Objects in Optical See-Through Augmented Reality https://paper.nweon.com/12060 Thu, 21 Apr 2022 05:55:28 +0000 https://paper.nweon.com/12060 PubDate: November 2021

Measuring the Perceived Three-Dimensional Location of Virtual Objects in Optical See-Through Augmented Reality最先出现在Nweon Paper

]]>
PubDate: November 2021

Teams: Mississippi State University;University of Nebraska–Lincoln

Writers: Farzana Alam Khan; Veera Venkata Ram Murali Krishna Rao Muvva; Dennis Wu; Mohammed Safayet Arefin; Nate Phillips; J. Edward Swan

PDF: Measuring the Perceived Three-Dimensional Location of Virtual Objects in Optical See-Through Augmented Reality

Abstract

For optical see-through augmented reality (AR), a new method for measuring the perceived three-dimensional location of virtual objects is presented, where participants verbally report a virtual object’s location relative to both a vertical and horizontal grid. The method is tested with a small (1.95 × 1.95 × 1.95 cm) virtual object at distances of 50 to 80 cm, viewed through a Microsoft HoloLens 1 st generation AR display. Two experiments examine two different virtual object designs, whether turning in a circle between reported object locations disrupts HoloLens tracking, and whether accuracy errors, including a rightward bias and underestimated depth, might be due to systematic errors that are restricted to a particular display. Turning in a circle did not disrupt HoloLens tracking, and testing with a second display did not suggest systematic errors restricted to a particular display. Instead, the experiments are consistent with the hypothesis that, when looking downwards at a horizontal plane, HoloLens 1 st generation displays exhibit a systematic rightward perceptual bias. Precision analysis suggests that the method could measure the perceived location of a virtual object within an accuracy of less than 1 mm.

Measuring the Perceived Three-Dimensional Location of Virtual Objects in Optical See-Through Augmented Reality最先出现在Nweon Paper

]]>
Walking Through Walls: The Effect of Collision-Based Feedback on Affordance Judgments in Augmented Reality https://paper.nweon.com/12028 Thu, 21 Apr 2022 04:34:43 +0000 https://paper.nweon.com/12028 PubDate: November 2021

Walking Through Walls: The Effect of Collision-Based Feedback on Affordance Judgments in Augmented Reality最先出现在Nweon Paper

]]>
PubDate: November 2021

Teams: University of Utah;Vanderbilt University

Writers: Holly C. Gagnon; Dun Na; Keith Heiner; Jeanine Stefanucci; Sarah Creem-Regehr; Bobby Bodenheimer

PDF: Walking Through Walls: The Effect of Collision-Based Feedback on Affordance Judgments in Augmented Reality

Abstract

Feedback about actions in augmented reality (AR) is limited and can be ambi due to the nature of interacting with virtual objects. AR devices also have a restricted field of view (FOV), limiting the amount of available visual information that can be used to perform an action or provide feedback during or after an action. We used the Microsoft HoloLens 1 to investigate whether perceptual-motor, collision-based outcome feedback calibrates judgments of whether one can pass through an aperture in AR. Additionally, we manipulated the amount of information available within the FOV by having participants view the aperture at two different distances. Feedback calibrated passing-through judgments at both distances but resulted in an overestimation of the just-passable aperture width. Moreover, the far viewing condition had more overestimation of just-passable aperture width than the near viewing condition.

Walking Through Walls: The Effect of Collision-Based Feedback on Affordance Judgments in Augmented Reality最先出现在Nweon Paper

]]>
Spatial Computing and Intuitive Interaction: Bringing Mixed Reality and Robotics Together https://paper.nweon.com/11843 Tue, 08 Mar 2022 23:34:25 +0000 https://paper.nweon.com/11843 PubDate: Feb 2022

Spatial Computing and Intuitive Interaction: Bringing Mixed Reality and Robotics Together最先出现在Nweon Paper

]]>
PubDate: Feb 2022

Teams: Microsoft Mixed Reality and AI Lab;ETH Zurich

Writers: Jeffrey Delmerico, Roi Poranne, Federica Bogo, Helen Oleynikova, Eric Vollenweider, Stelian Coros, Juan Nieto, Marc Pollefeys

PDF: Spatial Computing and Intuitive Interaction: Bringing Mixed Reality and Robotics Together

Abstract

Spatial computing – the ability of devices to be aware of their surroundings and to represent this digitally – offers novel capabilities in human-robot interaction. In particular, the combination of spatial computing and egocentric sensing on mixed reality devices enables them to capture and understand human actions and translate these to actions with spatial meaning, which offers exciting new possibilities for collaboration between humans and robots. This paper presents several human-robot systems that utilize these capabilities to enable novel robot use cases: mission planning for inspection, gesture-based control, and immersive teleoperation. These works demonstrate the power of mixed reality as a tool for human-robot interaction, and the potential of spatial computing and mixed reality to drive the future of human-robot interaction.

Spatial Computing and Intuitive Interaction: Bringing Mixed Reality and Robotics Together最先出现在Nweon Paper

]]>
Accuracy Evaluation of Touch Tasks in Commodity Virtual and Augmented Reality Head-Mounted Displays https://paper.nweon.com/11320 Thu, 14 Oct 2021 04:26:14 +0000 https://paper.nweon.com/11320 PubDate: Sep 2021

Accuracy Evaluation of Touch Tasks in Commodity Virtual and Augmented Reality Head-Mounted Displays最先出现在Nweon Paper

]]>
PubDate: Sep 2021

Teams: Coburg University of Applied Sciences and Arts;University of Primorska;Microsoft Research;University of Cambridge

Writers: Daniel Schneider, Verena Biener, Alexander Otte, Travis Gesslein, Philipp Gagel, Cuauhtli Campos, Klen Čopič Pucihar, Matjaž Kljun, Eyal Ofek, Michel Pahud, Per Ola Kristensson, Jens Grubert

PDF: Accuracy Evaluation of Touch Tasks in Commodity Virtual and Augmented Reality Head-Mounted Displays

Abstract

An increasing number of consumer-oriented head-mounted displays (HMD) for augmented and virtual reality (AR/VR) are capable of finger and hand tracking. We report on the accuracy of off-the-shelf VR and AR HMDs when used for touch-based tasks such as pointing or drawing. Specifically, we report on the finger tracking accuracy of the VR head-mounted displays Oculus Quest, Vive Pro and the Leap Motion controller, when attached to a VR HMD, as well as the finger tracking accuracy of the AR head-mounted displays Microsoft HoloLens 2 and Magic Leap. We present the results of two experiments in which we compare the accuracy for absolute and relative pointing tasks using both human participants and a robot. The results suggest that HTC Vive has a lower spatial accuracy than the Oculus Quest and Leap Motion and that the Microsoft HoloLens 2 provides higher spatial accuracy than Magic Leap One. These findings can serve as decision support for researchers and practitioners in choosing which systems to use in the future.

Accuracy Evaluation of Touch Tasks in Commodity Virtual and Augmented Reality Head-Mounted Displays最先出现在Nweon Paper

]]>
Rotation-constrained optical see-through headset calibration withbare-hand alignment https://paper.nweon.com/11288 Tue, 12 Oct 2021 03:01:19 +0000 https://paper.nweon.com/11288 PubDate: Aug 2021

Rotation-constrained optical see-through headset calibration withbare-hand alignment最先出现在Nweon Paper

]]>
PubDate: Aug 2021

Teams: Imperial College London;University of Pisa

Writers: Xue Hu, Ferdinando Rodriguez y Baena, Fabrizio Cutolo

PDF: Rotation-constrained optical see-through headset calibration withbare-hand alignment

Abstract

The inaccessibility of user-perceived reality remains an open issue in pursuing the accurate calibration of optical see-through (OST) head-mounted displays (HMDs). Manual user alignment is usually required to collect a set of virtual-to-real correspondences, so that a default or an offline display calibration can be updated to account for the user’s eye position(s). Current alignment-based calibration procedures usually require point-wise alignments between rendered image point(s) and associated physical landmark(s) of a target calibration tool. As each alignment can only provide one or a few correspondences, repeated alignments are required to ensure calibration quality.
This work presents an accurate and tool-less online OST calibration method to update an offline-calibrated eye-display model. The user’s bare hand is markerlessly tracked by a commercial RGBD camera anchored to the OST headset to generate a user-specific cursor for correspondence collection. The required alignment is object-wise, and can provide thousands of unordered corresponding points in tracked space. The collected correspondences are registered by a proposed rotation-constrained iterative closest point (rcICP) method to optimise the viewpoint-related calibration parameters. We implemented such a method for the Microsoft HoloLens 1. The resiliency of the proposed procedure to noisy data was evaluated through simulated tests and real experiments performed with an eye-replacement camera. According to the simulation test, the rcICP registration is robust against possible user-induced rotational misalignment. With a single alignment, our method achieves 8.81 arcmin (1.37 mm) positional error and 1.76 degree rotational error by camera-based tests in the arm-reach distance, and 10.79 arcmin (7.71 pixels) reprojection error by user tests.

Rotation-constrained optical see-through headset calibration withbare-hand alignment最先出现在Nweon Paper

]]>