Tri-view Morphing |
|||||||||
This paper presents an efficient image-based approach to navigate a scene based on only three wide-baseline uncalibrated images without the explicit use of a 3D model. After automatically recovering corresponding points between each pair of images, an accurate trifocal plane is extracted from the trifocal tensor of these three images. Next, based on a small number of feature marks from a friendly GUI, the correct dense disparity maps are obtained by using our trinocular-stereo algorithm. Employing the barycentric warping scheme with the computed disparity, we can generate an arbitrary novel view located in the trifocal plane. Furthermore, after self-calibration of the cameras, we correctly augment 3D objects into the virtual environment synthesized by the tri-view morphing algorithm. We also
demonstrate three applications of the tri-view morphing algorithm. The first one
is 4D video synthesis, which can be used to fill in the gap between a few
sparsely located video cameras to synthetically generate a video from a virtual
moving camera. This synthetic camera can be used to view the dynamic scene from
a novel view instead of the original static camera views. The second application
is multiple view morphing, where we can seamlessly fly through the scene over a
2D space constructed by more than three cameras. The last one is dynamic scene
synthesis using three still images, where several rigid objects may move in any
orientation or direction. After segmenting three reference frames into several
layers, the novel views in the dynamic scene can be generated by applying our
algorithm. Finally, the experiments are presented to illustrate that a series of
photo-realistic virtual views can be generated to fly through a virtual
environment covered by several static cameras. |
|||||||||
|
|||||||||
|
|||||||||
|
|||||||||
|
|||||||||
|