There are two main types of mixed reality systems, these are optical see-through mixed and video see-through devices. The main difference between them is how the user observes the real-world. In the case of the optical see-through mixed reality system, the device user observes the real-world directly, and in the case of the video see-through mixed reality system, they observe the real world projected from the LCD screens to the eye retina. As a result, in the first case, the vergence of human eyes corresponds to the point where sight is directed, however in the case of the video see-through system it is not true as the direction of human sight might not correspond to the fixed direction of device cameras. Moreover, these cameras might be shifted to the top or bottom of a mixed reality device or replaced with a single RGBD camera which might result in additional discomfort. So, when observing the real-world using the video see-through mixed reality system an additional image processing is required to lower the disagreement of the vergence between device cameras and user's eyes according to the data acquired from the device eye-tracking system. In the scope of the current research, authors present the results of research of lowering the vergence disagreement based on the rough depth map restoration and processing images from device cameras as a cloud of points. Various external camera setups and corresponding vergence correction results are compared.