I’m trying to acquire a point cloud and then capture an image (the capture would stop the point cloud acquisition session).
The point cloud acquisition seems easy. This is a well discussed topic: https://developer.apple.com/forums/thread/658109
Furthermore, it would seem possible to store camera pose information from the point cloud with the photo so that data could be used to find a pose. Later, I’d align the 3D point cloud to that 2d image (preferably not using depth images, interpolation, or NNs).
I am surprised, but seems really hard to find out the feasibility of such an approach.
Post
Replies
Boosts
Views
Activity
I am trying to create reflections and lighting in a fully virtual scene, and the way the documentation is written, the VirtualEnvironmentProbeComponent would be key. I cannot seem to get it to work, however. Has anyone got this figured out?