Apple's new Vision Pro headset is a completely new device in the Apple ecosystem. While the company initially provided details about the device during its 2-hour+ WWDC 2023 keynote, the company saved developer details for a later event, the Platforms State of the Union.
Also: Apple Vision Pro in stores February 2, pre-orders start January 19
Vision Pro's operating system, VisionOS, is designed for what Apple calls "spatial computing." The company separates this computing paradigm from the two we're most familiar with, desktop and mobile computing. The idea with spatial computing is that your work environment exists floating in front of you. We'll come back to this "spatial computing" buzzword in our 18th tip, later in the article.
The "shared space" is where apps float side by side. Think of it as multiple side-by-side windows, but instead of on a desktop, they're in midair.
Users can open one or more windows that exist as planes in space. They support traditional views and controls, but also support 3D content, which can live in a window with 2D content. In a CAD program, the object might be 3D, but the toolbar might be 2D, for example.
Beyond windows, apps can create three-dimensional volumes. These can contain objects and scenes. The key difference is that volumes can be moved around in 3D space and can be viewed from all angles. It's the difference between looking into a store's window display compared to walking around a car and peering in through front, back, and side windows.
Also: I tried Apple Vision Pro and it's far ahead of where I expected
For developers who want full immersion, it's possible to create a dedicated full space. This is like when a game takes over the entire screen, but in the VisionOS experience, that screen is fully immersive. Here, apps, windows, and volumes operate inside the fully immersive environment.
Now that you understand the virtual paradigms used by VisionOS, let's look at eighteen things developers need to know about developing for VisionOS.
Development is based on Swift UI, RealityKit, and ARKit, which are existing APIs that have been around for a while. Apple has extended these frameworks for VisionOS, adding support for the new hardware and full-space paradigm.
Developers will use SwiftUI and UIKit to build the user interface. RealityKit is used to display 3D content, animations, and visual effects. ARKit provides apps with an understanding of the real-world space around the user and make that understanding available to code within an app.
Even basic 2D apps ported over from iOS or iPad OS will float in space. Whether that space is a view of the room where the user is using the Vision Pro environment or some simulated environment that blocks the real world out, even traditional apps will "float" in 3D space.
Previously, Xcode developers could choose iPhone, iPad, and Mac as destinations (i.e., where the app would run). Now, developers can add VisionOS as a destination. As soon as the app is rebuilt, the new destination adds VisionOS features including resizable windows and the adaptive translucency features of VisionOS.
When they do, they'll get some highlight and 3D presence features from VisionOS. So while UIKit and Objective C-based apps may not be able to provide a fully immersive 3D experience, they will gain a native VisionOS look and feel, and be able to coexist reasonably seamlessly with more modern SwiftUI-based applications.
This allows developers to push panes and controls into 3D space, allowing certain interface elements to float in front of, or behind other elements. This can allow developers to bring attention to certain elements.
Foveation describes an image processing technique where certain areas of an image get more detail than other areas. With VisionOS, the Vision Pro uses eye tracking to render the area of the scene being looked at in very high resolution, but reduce the resolution in peripheral vision. This reduces processing time in areas where the user is not putting full attention. Developers don't need to code for this. It's built into the OS.
By default, objects floating in 3D space gain the lighting and shadow characteristics of the space where the user is wearing the headset. Developers can provide an image-based lighting asset if they want to customize how objects are lit in virtual space.
It uses plane estimation to identify flat surfaces in the real room. Scene reconstruction builds a dynamic 3D model of the space in the room that apps can interact with. Image anchoring allows 2D graphics to be locked into a location in the 3D space, making it appear to be part of the real world.
This provides apps with positioning data and joint mapping, so that gestures can more fully control the virtual experience. Accessibility features allow users to interact with eye movement, voice, and head movement in addition to hand movements.
Apple has partnered with Unity so Unity developers can directly target VisionOS from within Unity, allowing all of the Unity-based content to migrate into VisionOS-based apps without much conversion effort. This is actually pretty big because it allows developers with deep Unity experience to create Unity-based apps alongside VisionOS apps.
This is essentially an asset manager for 3D and virtual content. This also allows developers to create custom materials, test out shaders, integrate these assets into the Xcode development process, and preview on the Vision Pro.
This means that the room visuals and mapping is kept private. Cloud processing is not used for 3D mapping. All personal information and room spatial dynamics are managed entirely inside the Vision Pro device.
This allows you to get a feel for what your app will look like and test your app. The preview mode lets you see your layout in Xcode, while the simulator is a dedicated on-screen environment for testing overall app behavior. You can simulate gestures using a keyboard, trackpad, or game controller.
The Vision Pro extends Mac desktops into virtual space, which means you can have your Xcode development environment side-by-side with your Vision Pro app.
Apps, complete with in-app purchases, will be available to download and buy from the Vision Pro's dedicated app store. Additionally, Test Flight runs with Xcode and Vision OS as expected, so developers will be able to distribute betas of apps in exactly the same way as for iPhone and iPad.
The VisionOS SDK, updated Xcode, simulator, and Reality Composer Pro are available to developers. Apple has also set up Apple Vision Pro developer labs. Located in London, Munich, Shanghai, Singapore, Tokyo, and Cupertino, developers can make appointments to visit and test applications. For those who don't travel to Apple sites, developers can submit requests for Apple to evaluate and test app builds and provide feedback. Apple made no mention of turnaround time on these requests.
Apple has created an app store specifically for the Vision Pro. Developers can submit both native visionOS apps and modified iPad and iPhone apps to the Vision Pro store. This page on the Apple site provides detailed guidelines for submitting your app. One interesting restriction: Apple doesn't want you to describe your apps as augmented reality (AR), virtual reality (VR), extended reality (XR), or mixed reality (MR). Vision Pro appsmustbe described as "spatial computing" apps.
More information on developing for the Vision Pro and VisionOS is available on Apple's developer website.
Also: This ultraportable VR headset gave me a taste of Vision Pro at a fraction of the cost
So what do you think? Are you a developer? If so, are you planning on developing for the Vision Pro? Are you a user? Do you see an immediate use for this device or does the$3,500 price tag and goggle-like usage experience put you off? Please let us know in the comments below.
You can follow my day-to-day project updates on social media. Be sure to follow me on Twitter at @DavidGewirtz, on Facebook at Facebook.com/DavidGewirtz, on Instagram at Instagram.com/DavidGewirtz, and on YouTube at YouTube.com/DavidGewirtzTV.