WWDC25 Apple Vision Video Notes session #304 #297

(1) Started here: Explore video experiences for visionOS

https://developer.apple.com/videos/play/wwdc2025/304

  • APMP is packaging for all kinds of camera video
  • avconvert command is available
  • high motion scenes have automatic immersion decrease
    • detect high motion in settings (can be disabled)
  • ursa cine
    • 59 megapixels per eye, 90 fps
    • sharpness approaching human eye
    • 210 across and 180 up down
  • Apple Immersive Video utility app for previewing
  • framework read / write apple immersive programmatically (AVAssetWriter and CMTaggedBuffer)
  • Apple Spatial Audio format?

(2) Dive into: Learn about the Apple Projected Media Profile

https://developer.apple.com/videos/play/wwdc2025/297

  • VR180 stereo in mov/mp4 means it has to have both stereo metadata and projection hequ (half equirectangular)
  • APMP is available exported (they call this metadata … signaling?) as .mov or .mp4 from EOS VR Utility
  • CMTaggedBuffers available for recomposition of video an AVAssetWriterInput
  • Stereoscopic recommendations are
    • 4320×4320 per eye
    • 4k @ 30 FPS
    • 84 Mb/sec for 10-bit
      • 70Mb/sec for 8-bit
    • HEVC Main or HEVC Main 10
    • chroma 4:2:0
    • Color Rec. 709, P3-D65
    • does not specify if this is for flat stereo like iPhone Pro?
    • They recommend not exceeding 150 megabits per second peak
      • “Bitrate encoding settings are content dependent and should be chosen appropriately for your use case, but we recommend not exceeding 150 megabits per second peak.”
  • AVQT 2.0 tools, helps evaluate perceptional quality of half-equirect as well as bit-rate ladders
    • seems released from 2021 but updated for projection?
  • HLS tools recognize APMP
  • APAC for ambisonic sound
    • 1st-order is 4 mics/channels? 2nd order is 9, 3rd is 16
    • same file plays on all apple platforms (AirPods spatial sound?)