As far as I know, the working principle of Matterport as well as Kinect, is structured light scanning. It is quite a simple ranging method based on triangles with known baseline and two angles, but their accuracy degrades very fast as distance to object increases.
It has a lot to do with photogrammetry and Photoscan fundamentals. Here a structured infrared pattern is projected onto object surfaces playing a similar role of the object's texture we exploit in Photoscan. The relative orientation between camera(s) and projector is fixed and well known, as it is camera lens intrinsic parameters, so there is no need for camera calibration nor ground control to solve a single shot.
The advantages are:
- it is an active sensor (no need for external illumination to get 3D)
- is non dependent on object's texture
- 3D from one single point of view (just apparently)
- Fast! very fast!
The disadvantages are:
- Bad accuracy and poor resolution at long range
- poor color rendition
- frame stitching on the fly, not easy to access raw data to improve alignment or use image-picked ground control