I want to convert 2D Image coordinates to 3D world coordinates. I am using the ZED camera which is a stereo camera and the sdk shipped with it, provides the disparity map. Hence I have depth. The two cameras are parallel to each other. Although this problem is of trivial nature, I am unable to figure out the mathematics behind the conversion. I have the following information
1) I have the pixel information (i.e. the row and column number denoted by u and v) and the depth D at that pixel in meters.
2) I also know the focal length and cx, cy values of both the cameras
3) The distance between the two centers of projection B(baseline) is known.
I want to know how can one go from pixel(row, column, depth) in image to (X,Y,Z) in world coordinate.
Assume origin of world coordinate system as the central point between two cameras and vertically below on the ground. (The height at which camera is known as H).
Thank you.