Get camera pose information for point cloud calibration

I’m trying to acquire a point cloud and then capture an image (the capture would stop the point cloud acquisition session).

The point cloud acquisition seems easy. This is a well discussed topic: https://developer.apple.com/forums/thread/658109

Furthermore, it would seem possible to store camera pose information from the point cloud with the photo so that data could be used to find a pose. Later, I’d align the 3D point cloud to that 2d image (preferably not using depth images, interpolation, or NNs).

I am surprised, but seems really hard to find out the feasibility of such an approach.

Get camera pose information for point cloud calibration
 
 
Q