Return to search

Investigating Asymmetric Collaboration and Interaction in Immersive Environments

With the commercialization of virtual/augmented reality (VR/AR) devices, there is an increasing interest in combining immersive and non-immersive devices (e.g., desktop computers, mobile devices) for asymmetric collaborations.
While such asymmetric settings have been examined in social platforms, questions surrounding collaborative view dimensionalities in data-driven decision-making and interaction from non-immersive devices remain under-explored.
A crucial inquiry arises: although presenting a consistent 3D virtual world on both immersive and non-immersive platforms has been a common practice in social applications, does the same guideline apply to lay out data?
Or should data placement be optimized locally according to each device's display capacity?
To this effect, a user study was conducted to provide empirical insights into the user experience of asymmetric collaboration in data-driven decision-making.
The user study tested practical dimensionality combinations between PC and VR, resulting in three conditions: PC2D+VR2D, PC2D+VR3D, and PC3D+VR3D.
The results revealed a preference for PC2D+VR3D, and PC2D+VR2D led to the quickest task completion.
Similarly, mobile devices have become an inclusive alternative to head-worn displays in virtual reality (VR) environments, enhancing accessibility and allowing cross-device collaboration.
Object manipulation techniques in mobile Augmented Reality (AR) have been typically evaluated in table-top scale and we lack an understanding of how these techniques perform in room-scale environments.
Two studies were conducted to analyze object translation tasks, each with 30 participants, to investigate how different techniques impact usability and performance for room-scale mobile VR object translations.
Results indicated that the Joystick technique, which allowed translation in relation to the user's perspective, was the fastest and most preferred, without difference in precision.
These findings provide insight for designing collaborative, asymmetric VR environments. / Master of Science / With the commercialization of virtual/augmented reality (VR/AR) devices, there is an increasing interest in combining immersive and non-immersive devices (e.g., desktop computers, mobile devices) for collaborations across different devices.
While such asymmetric settings have been examined in social platforms, questions surrounding collaborative view differences in 2D views or 3D views affect data-driven decision-making and interaction remain under-explored.
A crucial inquiry arises: although presenting a consistent 3D virtual world on both immersive and non-immersive platforms has been a common practice in social applications, does the same guideline apply to lay out data?
Or should data placement be optimized on each device according to each device's display capacity?
To this effect, a user study was conducted to provide insights into the user experience of collaboration across different devices in data-driven decision-making.
The user study tested different combinations of 2D and 3D layouts between PC and VR, resulting in three conditions: PC2D+VR2D, PC2D+VR3D, and PC3D+VR3D.
The results revealed a preference for PC2D+VR3D, and PC2D+VR2D led to the quickest task completion.
Similarly, mobile devices have become an inclusive alternative to head-worn displays in virtual reality (VR) environments, enhancing accessibility and allowing cross-device collaboration.
Object manipulation techniques in mobile Augmented Reality (AR) have been typically evaluated in table-top scale and we lack an understanding of how these techniques perform in room-scale environments.
Two studies were conducted to analyze object translation tasks, each with 30 participants, to investigate how different techniques impact usability and performance for room-scale mobile VR object translations.
Results indicated that the Joystick technique, which allowed translation in relation to the user's perspective, was the fastest and most preferred, without difference in precision.
These findings provide insight for designing collaborative, asymmetric VR environments.

Identiferoai:union.ndltd.org:VTETD/oai:vtechworks.lib.vt.edu:10919/117630
Date23 January 2024
CreatorsEnriquez, Daniel
ContributorsComputer Science and Applications, Yang, Yalong, Lee, Sang Won, North, Christopher L., David-John, Brendan Matthew
PublisherVirginia Tech
Source SetsVirginia Tech Theses and Dissertation
LanguageEnglish
Detected LanguageEnglish
TypeThesis
FormatETD, application/pdf
RightsIn Copyright, http://rightsstatements.org/vocab/InC/1.0/

Page generated in 0.0024 seconds