I have the same issue. But If I lauch the app with Debugger Mode (XCode) it always return the same issue above "[MTLDebugDevice validateNewBufferArgs:options:]:467: failed assertion `Cannot create buffer of zero length.'". But if I lauch without Debugger Mode (not include XCode) it's okay but in that case the shape is not correctly with my wishes, of course it doesn't crash my app.
So Anyone can tell me
Why in Debugger Mode, it will broken my app but in Release Mode it doesn't?
In Release Mode, Why Sometimes the Sceneview can show my shape with "cross line" and sometimes it will not?
I really thank and appreciate any advice on my issue.
Post
Replies
Boosts
Views
Activity
Thanks @darknoon for your help.
"Sorry, I'm not really following, each value in the sceneDepth.depthMap contains depth data (distance in meters from the rear camera). Is that somehow different from what you describe as "depth data (distance)"?"-- Thank for your answer, I am new in swift iOS, so it loss many time for get value from CVpixelbuffer (by pointer)
-- I have two question, can you help me for issue
1/ I tried for save CVPixelBuffer to image 16 bit depth. It's meaning, I have to convert 32 bit value of CVpixelbuffer for image 16 bit but I can't find any function for that (I have to use python for convert)
I used CVpixelbuffer -- CIImage -- UIImage (1) -- Data (pngData) -- UIImage(2) -- UIImagewritetoPhotosAlbum
At (1), I don't no why i can't save to image because I think (1) and (**2) is the same type of data
2/ The depth value is short, 192x256 (49152) point. When i convert depth to point cloud (it very very hard to see) but when i combine with rgb color (image from 192x256 to 8 bit), it look like OKE. The question, "depth" is a combination of depth data and Rgb data, it's right
Thank you very much.
Have you got depth data (distance) with ARFrame yet?