TL;DR: Snapdragon Spaces™ XR Developer Platform is Qualcomm Technologies’ end-to-end platform for building XR experiences on compatible headsets powered by Android. The platform’s Dual Render Fusion feature enables smartphone screens to become a primary input for AR applications, while AR glasses act as a secondary augmented display. The dual display capability allows developers and users to run new or existing apps in 2D on the smartphone screen while showing additional content in 3D in augmented reality.
Snapdragon Spaces™ XR Developer Platform is Qualcomm Technologies’ end-to-end platform for building XR experiences on compatible headsets powered by Android. Recently, Snapdragon Spaces introduced Dual Render Fusion — a new feature in the SDK designed to help developers transform their 2D mobile Android applications into spatial 3D experiences with little prior knowledge required.
What is Dual Render Fusion?
Snapdragon Spaces Dual Render Fusion enables smartphone screens to become a primary input for AR applications, while AR glasses act as a secondary augmented display. The dual display capability allows developers and users to run new or existing apps in 2D on the smartphone screen while showing additional content in 3D in augmented reality. In practical terms, a smartphone acts as a controller for AR experiences, letting users select what they want to see in AR using familiar mobile UI and gestures. Imagine you are using your go-to maps app for sightseeing. With Dual Render Fusion, you can use the phone as usual to browse the map and at the same time, see a 3D reconstruction of historical places in AR.
Watch the following video to see examples of end-user experiences that Dual Render Fusion can provide:
Why use Dual Render Fusion in your AR experiences?
The feature makes it easier for you to extend your 2D mobile Android apps into 3D spatial experiences without creating a new spatial UI. It’s also the first time in the XR industry that AR development tools have been made available to combine multi-modal input with simultaneous rendering to smartphones and AR glasses. Using Dual Render Fusion, developers with Unity experience can easily add an AR layer to their existing app using just a few extra lines of code. The feature gives more control over app behavior in the 3D space, significantly lowering the entry barrier to AR. But that’s not all — while using the feature, you have the option to utilize all available inputs enabled with the Snapdragon Spaces SDK, including Hand Tracking, Spatial Mapping and Meshing, Plane Detection, and Image Tracking, or go all in utilizing the convenience of the mobile touch screen for all input.
Why is Dual Render Fusion significant? Why are we at an inflection point?
It takes a lot of learning for developers to break into mobile AR and even more to rethink what you already know, to then apply spatial design principles to headworn AR. The same applies to your end users who need to get familiar with new spatial UX/UI and input. Enabling the majority of developers to create applications which are accessible to smartphone users will unlock the great potential of smartphone-based AR. Snapdragon Spaces team members have been working hard to reimagine smartphone AR’s status quo and take a leap forward to fuse the phone with AR glasses. The Dual Render Fusion feature allows just that — to blend the simplicity and familiarity of the smartphone touch screen for input while leveraging the best of augmented reality. Dual Render Fusion unlocks smartphone-powered AR to its full potential and opens it up to millions of mobile developers like you.
Download and Build your AR app now
Dual Render Fusion is now available in beta as an optional add-on package for Snapdragon Spaces SDK for Unity version 0.13.0 and above, so be sure to browse the Documentation to learn more.
Snapdragon branded products are products of Qualcomm Technologies, Inc. and/or its subsidiaries.
This article was previously published on proandroiddev.com