🤖 AI Summary
Existing smartphone-based augmented reality (AR) systems predominantly rely on a single camera—either front- or rear-facing—resulting in limited field-of-view, insufficient screen real estate, and diminished practicality and immersion.
Method: This paper proposes a dual-camera collaborative AR interaction framework, introducing the first systematic design principles for dual-camera AR: balancing real-world behavioral relevance with multimodal feedback quality. It mitigates user spatial disorientation through synchronized video streaming and adaptive camera switching. Implemented on commercial smartphones, the framework integrates real-time front/rear camera video, visual and haptic feedback, context awareness, and user behavior modeling.
Contribution/Results: A qualitative study with 30 participants yielded two reusable design guidelines that significantly enhance interaction naturalness and spatial orientation stability. The work provides both theoretical foundations and an implementable paradigm to advance the practicality and widespread adoption of dual-camera AR systems.
📝 Abstract
Off-the-shelf smartphone-based AR systems typically use a single front-facing or rear-facing camera, which restricts user interactions to a narrow field of view and small screen size, thus reducing their practicality. We present extit{Cam-2-Cam}, an interaction concept implemented in three smartphone-based AR applications with interactions that span both cameras. Results from our qualitative analysis conducted on 30 participants presented two major design lessons that explore the interaction space of smartphone AR while maintaining critical AR interface attributes like embodiment and immersion: (1) extit{Balancing Contextual Relevance and Feedback Quality} serves to outline a delicate balance between implementing familiar interactions people do in the real world and the quality of multimodal AR responses and (2) extit{Preventing Disorientation using Simultaneous Capture and Alternating Cameras} which details how to prevent disorientation during AR interactions using the two distinct camera techniques we implemented in the paper. Additionally, we consider observed user assumptions or natural tendencies to inform future implementations of dual-camera setups for smartphone-based AR. We envision our design lessons as an initial pioneering step toward expanding the interaction space of smartphone-based AR, potentially driving broader adoption and overcoming limitations of single-camera AR.