“…Specifically, a virtual image of viewpoint u can be synthesized at decoder via depth-image-based rendering (DIBR) [1] using texture and depth maps of left and right reference views vL and vR, where vL ≤ u ≤ vR. Many practical challenges in interactive free viewpoint systems have already been addressed in works such as [2,3,4,5,6,7,8,9], which propose novel coding and/or streaming strategies at the server to optimize the tradeoff between transmission rate and visual distortion, while providing the promised viewnavigation ability at the client. As an example, in order to deal with network delay, the system in [3] transmits a subset of camera views enabling synthesis of a window W of possible virtual views (with the window size proportional to the round trip time (RTT) between server and client), so that the client can freely navigate in the 3D scene.…”