Post not yet marked as solved
When I open a volumetric window, it appears slightly to the right of its parent (SwiftUI window) is this normal (it's not centered on the horizontal axis)?
I was expecting the center (x axis) of the volumetric window to be aligned with the center of the parent (x axis of the SwiftUI window).
Is this a bug in the simulator, or is it an intentional choice by Apple? (Some of Apple's documentations show this offset, which leads me to think that it is a deliberate choice... However, I don't understand the reason of this design decision...).
Post not yet marked as solved
Are there any technical recommendations regarding the display of a 3D scene in VisionOS?
I am particularly looking to know:
the maximum recommended number of polygons (e.g., display within a volume, throughout the entire space...)
the maximum recommended resolution of an HDR/EXR image used as a lighting source (e.g., 2K?)
the maximum recommended resolution of a png or jpeg/equirectangular texture used as a 360 background
Has anyone equipped with a headset conducted tests?
Thank you!
Post not yet marked as solved
I do not encounter any problems during the upload which seems to proceed correctly (success message at the end)...
However, I cannot view the builds from the interface (App Store Connect).
I've been testing different options (re-build, test on another computer, cache cleanup...) for several hours without success ;-(
@Apple: I need urgent help please
Post not yet marked as solved
What is the correct approach to save the image (eg. CVPixelBuffer to png) obtained after calling the captureHighResolutionFrame method?
"ARKit captures pixel buffers in a full-range planar YCbCr format (also known as YUV) format according to the ITU R. 601-4 standard"
Should I change the color space of the image (ycbcr to rgb using Metal)?