(1) Started here: Explore video experiences for visionOS
https://developer.apple.com/videos/play/wwdc2025/304
- APMP is packaging for all kinds of camera video
- avconvert command is available
- high motion scenes have automatic immersion decrease
- detect high motion in settings (can be disabled)
- ursa cine
- 59 megapixels per eye, 90 fps
- sharpness approaching human eye
- 210 across and 180 up down
- Apple Immersive Video utility app for previewing
- framework read / write apple immersive programmatically (AVAssetWriter and CMTaggedBuffer)
- Apple Spatial Audio format?
(2) Dive into: Learn about the Apple Projected Media Profile
https://developer.apple.com/videos/play/wwdc2025/297
- VR180 stereo in mov/mp4 means it has to have both stereo metadata and projection hequ (half equirectangular)
- APMP is available exported (they call this metadata … signaling?) as .mov or .mp4 from EOS VR Utility
- CMTaggedBuffers available for recomposition of video an AVAssetWriterInput
- Stereoscopic recommendations are
- 4320×4320 per eye
- 4k @ 30 FPS
- 84 Mb/sec for 10-bit
- 70Mb/sec for 8-bit
- HEVC Main or HEVC Main 10
- chroma 4:2:0
- Color Rec. 709, P3-D65
- does not specify if this is for flat stereo like iPhone Pro?
- They recommend not exceeding 150 megabits per second peak
- “Bitrate encoding settings are content dependent and should be chosen appropriately for your use case, but we recommend not exceeding 150 megabits per second peak.”
- AVQT 2.0 tools, helps evaluate perceptional quality of half-equirect as well as bit-rate ladders
- seems released from 2021 but updated for projection?
- HLS tools recognize APMP
- APAC for ambisonic sound
- 1st-order is 4 mics/channels? 2nd order is 9, 3rd is 16
- same file plays on all apple platforms (AirPods spatial sound?)