Abstract
In free viewpoint video (FVV) framework, a large number of viewpoints from limited number of views are generated to reduce the amount of transmitting, receiving and processing video data significantly. To generate a virtual view, the disparity among adjacent views or temporal correlation between different frames of the intermediate views are normally exploited. Those techniques may concern poor rendering quality by missing pixel values (i.e. creating holes) due to the occluded region, rounding error and disparity discontinuity. To address these problems recent techniques use inpainting, however, they still suffer quality degradation due to lack of spatial correlation on the foreground-background boundary areas. The background updating techniques with Gaussian mixture based modelling (GMM) can improve quality in some occluded areas, however, due to the dependencies on warping of background image and spatial correlation they still suffer quality degradation. In this paper, we propose a view synthesized prediction using Gaussian model (VSPGM) technique using the number of GMM models rather than the background image to identify background/foreground pixels. The missing pixels of background and foreground are recovered from the background pixel and the weighted average of warped and foreground model pixels respectively. The experimental results show that the proposed approach provides 0.50~2.14dB PSNR improved synthesized view compared with the state-of-the-art methods. To verify the effectiveness of the proposed synthesized view, we use it as a reference frame with immediate previous frame of current view in the motion estimation for multi-view video coding (MVC). The experimental results confirm that the proposed technique is able to improve PSNR by 0.17 to 1.00dB compared to the conventional three reference frames.
Original language | English |
---|---|
Title of host publication | Proceedings of the 2016 international conference on digital image computing: techniques and applications (DICTA) |
Editors | Alan Wee-Chung, Brian Lovell, Clinton Fookes, Jun Zhou, Yongsheng Gao, Michael Blumenstein, Zhiyong Wang |
Place of Publication | United States |
Publisher | IEEE, Institute of Electrical and Electronics Engineers |
Pages | 1-8 |
Number of pages | 8 |
ISBN (Electronic) | 9781509028962 |
ISBN (Print) | 9781509028979 (Print on demand) |
DOIs | |
Publication status | Published - 2016 |
Event | 2016 International Conference on Digital Image Computing: Techniques and Applications (DICTA) - Mantra on View Hotel, Surfer's Paradise, Gold Coast, Australia Duration: 30 Nov 2016 → 02 Dec 2016 https://web.archive.org/web/20161019111726/http://dicta2016.dictaconference.org/ (Conference website) http://dicta2016.dictaconference.org/program.html (Conference program) http://dicta2016.dictaconference.org/pdf/DICTA2016CFP.pdf (Call for papers) |
Conference
Conference | 2016 International Conference on Digital Image Computing: Techniques and Applications (DICTA) |
---|---|
Country/Territory | Australia |
City | Gold Coast |
Period | 30/11/16 → 02/12/16 |
Other | The International Conference on Digital Image Computing: Techniques and Applications (DICTA) is the main Australian Conference on computer vision, image processing, pattern recognition, and related areas. DICTA was established in 1991 as the premier conference of the Australian Pattern Recognition Society (APRS). |
Internet address |
|