You searched for subject:(View synthesis)
.
Showing records 1 – 21 of
21 total matches.
No search limiters apply to these results.
1.
Rongsirigul, Thiwat.
The Fast-NVS: Novel View Synthesis Using Accelerated View-dependent Texture Mapping for Stereoscopic HMD : The Fast-NVS : 自由視点ステレオ画像生成のための視点依存テクスチャマッピングの高速化; The Fast-NVS : ジユウ シテン ステレオ ガゾウ セイセイ ノ タメ ノ シテン イゾン テクスチャ マッピング ノ コウソクカ.
Degree: Nara Institute of Science and Technology / 奈良先端科学技術大学院大学
URL: http://hdl.handle.net/10061/12165
Subjects/Keywords: Novel view synthesis
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Rongsirigul, T. (n.d.). The Fast-NVS: Novel View Synthesis Using Accelerated View-dependent Texture Mapping for Stereoscopic HMD : The Fast-NVS : 自由視点ステレオ画像生成のための視点依存テクスチャマッピングの高速化; The Fast-NVS : ジユウ シテン ステレオ ガゾウ セイセイ ノ タメ ノ シテン イゾン テクスチャ マッピング ノ コウソクカ. (Thesis). Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Retrieved from http://hdl.handle.net/10061/12165
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Rongsirigul, Thiwat. “The Fast-NVS: Novel View Synthesis Using Accelerated View-dependent Texture Mapping for Stereoscopic HMD : The Fast-NVS : 自由視点ステレオ画像生成のための視点依存テクスチャマッピングの高速化; The Fast-NVS : ジユウ シテン ステレオ ガゾウ セイセイ ノ タメ ノ シテン イゾン テクスチャ マッピング ノ コウソクカ.” Thesis, Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Accessed December 14, 2019.
http://hdl.handle.net/10061/12165.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Rongsirigul, Thiwat. “The Fast-NVS: Novel View Synthesis Using Accelerated View-dependent Texture Mapping for Stereoscopic HMD : The Fast-NVS : 自由視点ステレオ画像生成のための視点依存テクスチャマッピングの高速化; The Fast-NVS : ジユウ シテン ステレオ ガゾウ セイセイ ノ タメ ノ シテン イゾン テクスチャ マッピング ノ コウソクカ.” Web. 14 Dec 2019.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Vancouver:
Rongsirigul T. The Fast-NVS: Novel View Synthesis Using Accelerated View-dependent Texture Mapping for Stereoscopic HMD : The Fast-NVS : 自由視点ステレオ画像生成のための視点依存テクスチャマッピングの高速化; The Fast-NVS : ジユウ シテン ステレオ ガゾウ セイセイ ノ タメ ノ シテン イゾン テクスチャ マッピング ノ コウソクカ. [Internet] [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; [cited 2019 Dec 14].
Available from: http://hdl.handle.net/10061/12165.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.
Council of Science Editors:
Rongsirigul T. The Fast-NVS: Novel View Synthesis Using Accelerated View-dependent Texture Mapping for Stereoscopic HMD : The Fast-NVS : 自由視点ステレオ画像生成のための視点依存テクスチャマッピングの高速化; The Fast-NVS : ジユウ シテン ステレオ ガゾウ セイセイ ノ タメ ノ シテン イゾン テクスチャ マッピング ノ コウソクカ. [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; Available from: http://hdl.handle.net/10061/12165
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.
2.
FABIAN LORENZO DAYRIT.
Novel View Synthesis and Augmented Reality for Assisting Human Action Learning : 動作学習支援のための自由視点映像生成と拡張現実感; ドウサ ガクシュウ シエン ノ タメ ノ ジユウ シテン エイゾウ セイセイ ト カクチョウ ゲンジツカン.
Degree: 博士(工学), 2017, Nara Institute of Science and Technology / 奈良先端科学技術大学院大学
URL: http://hdl.handle.net/10061/11683
Subjects/Keywords: novel view synthesis
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
DAYRIT, F. L. (2017). Novel View Synthesis and Augmented Reality for Assisting Human Action Learning : 動作学習支援のための自由視点映像生成と拡張現実感; ドウサ ガクシュウ シエン ノ タメ ノ ジユウ シテン エイゾウ セイセイ ト カクチョウ ゲンジツカン. (Thesis). Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Retrieved from http://hdl.handle.net/10061/11683
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
DAYRIT, FABIAN LORENZO. “Novel View Synthesis and Augmented Reality for Assisting Human Action Learning : 動作学習支援のための自由視点映像生成と拡張現実感; ドウサ ガクシュウ シエン ノ タメ ノ ジユウ シテン エイゾウ セイセイ ト カクチョウ ゲンジツカン.” 2017. Thesis, Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Accessed December 14, 2019.
http://hdl.handle.net/10061/11683.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
DAYRIT, FABIAN LORENZO. “Novel View Synthesis and Augmented Reality for Assisting Human Action Learning : 動作学習支援のための自由視点映像生成と拡張現実感; ドウサ ガクシュウ シエン ノ タメ ノ ジユウ シテン エイゾウ セイセイ ト カクチョウ ゲンジツカン.” 2017. Web. 14 Dec 2019.
Vancouver:
DAYRIT FL. Novel View Synthesis and Augmented Reality for Assisting Human Action Learning : 動作学習支援のための自由視点映像生成と拡張現実感; ドウサ ガクシュウ シエン ノ タメ ノ ジユウ シテン エイゾウ セイセイ ト カクチョウ ゲンジツカン. [Internet] [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; 2017. [cited 2019 Dec 14].
Available from: http://hdl.handle.net/10061/11683.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
DAYRIT FL. Novel View Synthesis and Augmented Reality for Assisting Human Action Learning : 動作学習支援のための自由視点映像生成と拡張現実感; ドウサ ガクシュウ シエン ノ タメ ノ ジユウ シテン エイゾウ セイセイ ト カクチョウ ゲンジツカン. [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; 2017. Available from: http://hdl.handle.net/10061/11683
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Université Catholique de Louvain
3.
Verleysen, Cédric.
3D estimation and view synthesis in wide-baseline stereo.
Degree: 2015, Université Catholique de Louvain
URL: http://hdl.handle.net/2078.1/167678
► Nowadays, when viewing a pre-recorded video content, the spectator's viewpoint is restricted to one of the cameras that have recorded the scene. To improve the…
(more)
▼ Nowadays, when viewing a pre-recorded video content, the spectator's viewpoint is restricted to one of the cameras that have recorded the scene. To improve the viewer's experience, the next generation of video content aims at enabling him/her to interactively define his/her viewpoint. This domain is known as view synthesis and consists in the interpolation of images that would be seen from a different viewpoint than the ones captured by real cameras. However, the current solutions require that the real cameras share very similar viewpoints, which limits the range of synthesized views. To circumvent this issue, this thesis focuses on the view interpolation when only two real cameras observe the scene from very different viewpoints. This minimalist and challenging camera configuration, called wide-baseline stereo, makes the view synthesis and its underlying 3D estimation problem ill-posed, i.e., multiple reconstructed views are possible. This thesis proposes three new priors to address the problem. The first contribution proposes an energy minimization framework to favor the preservation of the order of the elements in the scene while changing the viewpoint. The second contribution assumes the piecewise-planarity of the scene's background and approximates its 3D by a set of 3D planes under sparsity and smoothness constraints. The last contribution considers the view synthesis of dynamic foreground objects. It first learns a prior about plausible 2D silhouettes of the object based on non-linear dimensionality reduction of their shape descriptors. Afterwards, it constrains the interpolated views to be consistent with this shape prior.
(FSA - Sciences de l'ingénieur) – UCL, 2015
Advisors/Committee Members: UCL - SST/ICTM/ELEN - Pôle en ingénierie électrique, UCL - Ecole Polytechnique de Louvain, De Vleeschouwer, Christophe, Jacques, Laurent, Van Gool, Luc, Frossard, Pascal, Lafruit, Gauthier, Raskin, Jean-Pierre.
Subjects/Keywords: View interpolation; Free-viewpoint; Image-based rendering; Model-based rendering; View synthesis; Wide-baseline
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Verleysen, C. (2015). 3D estimation and view synthesis in wide-baseline stereo. (Thesis). Université Catholique de Louvain. Retrieved from http://hdl.handle.net/2078.1/167678
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Verleysen, Cédric. “3D estimation and view synthesis in wide-baseline stereo.” 2015. Thesis, Université Catholique de Louvain. Accessed December 14, 2019.
http://hdl.handle.net/2078.1/167678.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Verleysen, Cédric. “3D estimation and view synthesis in wide-baseline stereo.” 2015. Web. 14 Dec 2019.
Vancouver:
Verleysen C. 3D estimation and view synthesis in wide-baseline stereo. [Internet] [Thesis]. Université Catholique de Louvain; 2015. [cited 2019 Dec 14].
Available from: http://hdl.handle.net/2078.1/167678.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Verleysen C. 3D estimation and view synthesis in wide-baseline stereo. [Thesis]. Université Catholique de Louvain; 2015. Available from: http://hdl.handle.net/2078.1/167678
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Heriot-Watt University
4.
Isgro, Francesco.
Geometric methods for video sequence analysis and applications.
Degree: PhD, 2001, Heriot-Watt University
URL: http://hdl.handle.net/10399/495
Subjects/Keywords: 621.3994; Computer vision; View synthesis
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Isgro, F. (2001). Geometric methods for video sequence analysis and applications. (Doctoral Dissertation). Heriot-Watt University. Retrieved from http://hdl.handle.net/10399/495
Chicago Manual of Style (16th Edition):
Isgro, Francesco. “Geometric methods for video sequence analysis and applications.” 2001. Doctoral Dissertation, Heriot-Watt University. Accessed December 14, 2019.
http://hdl.handle.net/10399/495.
MLA Handbook (7th Edition):
Isgro, Francesco. “Geometric methods for video sequence analysis and applications.” 2001. Web. 14 Dec 2019.
Vancouver:
Isgro F. Geometric methods for video sequence analysis and applications. [Internet] [Doctoral dissertation]. Heriot-Watt University; 2001. [cited 2019 Dec 14].
Available from: http://hdl.handle.net/10399/495.
Council of Science Editors:
Isgro F. Geometric methods for video sequence analysis and applications. [Doctoral Dissertation]. Heriot-Watt University; 2001. Available from: http://hdl.handle.net/10399/495

University of Ottawa
5.
Pouladzadeh, Parvaneh.
Design and Implementation of Video View Synthesis for the Cloud
.
Degree: 2017, University of Ottawa
URL: http://hdl.handle.net/10393/37048
► In multi-view video applications, view synthesis is a computationally intensive task that needs to be done correctly and efficiently in order to deliver a seamless…
(more)
▼ In multi-view video applications, view synthesis is a computationally intensive task that needs to be done correctly and efficiently in order to deliver a seamless user experience. In order to provide fast and efficient view synthesis, in this thesis, we present a cloud-based implementation that will be especially beneficial to mobile users whose devices may not be powerful enough for high quality view synthesis. Our proposed implementation balances the view synthesis algorithm’s components across multiple threads and utilizes the computational capacity of modern CPUs for faster and higher quality view synthesis. For arbitrary view generation, we utilize the depth map of the scene from the cameras’ viewpoint and estimate the depth information conceived from the virtual camera. The estimated depth is then used in a backward direction to warp the cameras’ image onto the virtual view. Finally, we use a depth-aided inpainting strategy for the rendering step to reduce the effect of disocclusion regions (holes) and to paint the missing pixels. For our cloud implementation, we employed an automatic scaling feature to offer elasticity in
order to adapt the service load according to the fluctuating user demands. Our performance results using 4 multi-view videos over 2 different scenarios show that our proposed system achieves average improvement of 3x speedup, 87% efficiency, and 90% CPU utilization for the parallelizable parts of the algorithm.
Subjects/Keywords: View synthesis;
Cloud computing;
Elasticity;
Multi-threading;
Warping;
Homography
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Pouladzadeh, P. (2017). Design and Implementation of Video View Synthesis for the Cloud
. (Thesis). University of Ottawa. Retrieved from http://hdl.handle.net/10393/37048
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Pouladzadeh, Parvaneh. “Design and Implementation of Video View Synthesis for the Cloud
.” 2017. Thesis, University of Ottawa. Accessed December 14, 2019.
http://hdl.handle.net/10393/37048.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Pouladzadeh, Parvaneh. “Design and Implementation of Video View Synthesis for the Cloud
.” 2017. Web. 14 Dec 2019.
Vancouver:
Pouladzadeh P. Design and Implementation of Video View Synthesis for the Cloud
. [Internet] [Thesis]. University of Ottawa; 2017. [cited 2019 Dec 14].
Available from: http://hdl.handle.net/10393/37048.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Pouladzadeh P. Design and Implementation of Video View Synthesis for the Cloud
. [Thesis]. University of Ottawa; 2017. Available from: http://hdl.handle.net/10393/37048
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

University of Illinois – Urbana-Champaign
6.
Tsoi, Ka Wai.
Improve OpenMVG and create a novel algorithm for novel view synthesis from point clouds.
Degree: MS, Computer Science, 2016, University of Illinois – Urbana-Champaign
URL: http://hdl.handle.net/2142/90821
► This thesis presents work to improve open source 3D reconstruction software OpenMVG and to create a novel algorithm to render photorealistic images from new views…
(more)
▼ This thesis presents work to improve open source 3D reconstruction software OpenMVG and to create a novel algorithm to render photorealistic images from new views given a photo collection and 3D point cloud.
First, the original OpenMVG is parallelized using GPU and its data structure is optimized. Moreover, we integrated the MatchMiner algorithm into OpenMVG to further improve its efficiency. Last but not least, an initial pair selection formulation and a default focal length setting are introduced and implemented to automize OpenMVG. Then 3D sparse point clouds of construction sites are reconstructed by performing Structure-from-Motion (SfM) with the improved version of OpenMVG and source images (images that are used in SfM) are calibrated and registered to point clouds. Furukawa's Patch-based Multi-
view Stereo(PMVS) algorithm is used to reconstruct dense point clouds using calibrated cameras as inputs. With known depth values of 3D points in the dense point cloud, we estimate depth maps of source images using optimization similar to Levin's colorization algorithm. For a novel
view of the point cloud, we find source images that share some common elements of the construction site that are also visible to the novel
view. Then we warp depth maps of these candidate images to the novel
view. We estimate a depth map and label pixels for the novel
view by solving a multi-label Markov Random Field (MRF) optimization problem using graph-cuts. We introduce a novel energy minimization formulation exploits both 2D and 3D information. Finally, a photorealistic image of the novel
view is rendered by copying pixel colors from selected candidate source images using pixel labels computed with graph-cuts.
We experimentally validate our approach on several challenging viewing angles of a point cloud model of a complicate construction site. The rendered results show high photo-realistic
synthesis quality in planar scenes.
Advisors/Committee Members: Hoiem, Derek W (advisor).
Subjects/Keywords: Computer Vision; Computer Science; Structure-from-Movtion; Novel View Synthesis
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Tsoi, K. W. (2016). Improve OpenMVG and create a novel algorithm for novel view synthesis from point clouds. (Thesis). University of Illinois – Urbana-Champaign. Retrieved from http://hdl.handle.net/2142/90821
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Tsoi, Ka Wai. “Improve OpenMVG and create a novel algorithm for novel view synthesis from point clouds.” 2016. Thesis, University of Illinois – Urbana-Champaign. Accessed December 14, 2019.
http://hdl.handle.net/2142/90821.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Tsoi, Ka Wai. “Improve OpenMVG and create a novel algorithm for novel view synthesis from point clouds.” 2016. Web. 14 Dec 2019.
Vancouver:
Tsoi KW. Improve OpenMVG and create a novel algorithm for novel view synthesis from point clouds. [Internet] [Thesis]. University of Illinois – Urbana-Champaign; 2016. [cited 2019 Dec 14].
Available from: http://hdl.handle.net/2142/90821.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Tsoi KW. Improve OpenMVG and create a novel algorithm for novel view synthesis from point clouds. [Thesis]. University of Illinois – Urbana-Champaign; 2016. Available from: http://hdl.handle.net/2142/90821
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Universidade do Rio Grande do Sul
7.
Oliveira, Adriano Quilião de.
Síntese de vistas em depht-image-based rendering (DIBR).
Degree: 2016, Universidade do Rio Grande do Sul
URL: http://hdl.handle.net/10183/143155
► Esta dissertação investiga soluções para o problema genérico de geração de vistas sintéticas a partir de um conjunto de imagens utilizando a abordagem Depth-Image-Based Rendering.…
(more)
▼ Esta dissertação investiga soluções para o problema genérico de geração de vistas sintéticas a partir de um conjunto de imagens utilizando a abordagem Depth-Image-Based Rendering. Essa abordagem utiliza um formato compacto para a representação de imagens 3D, composto basicamente por duas imagens, uma colorida para a vista de referência e outra em tons de cinza com a correspondência de disparidade para cada pixel. Soluções para esse problema beneficiam aplicações como Free Viewpoint Television. O maior desafio é o preenchimento de regiões sem informação de projeção considerando o novo ponto de vista, genericamente denominados holes, além de outros artefatos como cracks e ghosts que ocorrem por oclusões e erros no mapa de disparidade. Nesta dissertação apresentamos técnicas para remoção e tratamento de cada uma das classes de potenciais artefatos. O conjunto de métodos propostos apresenta melhores resultados quando comparado com o atual estado da arte em geração de vistas sintéticas com o modelo DIBR para o conjunto de dados Middlebury, considerando-se as métricas SSIM e PSNR.
This dissertation investigates solutions to the general problem of generating synthetic views from a set of images using the Depth-Image-Based Rendering approach. This approach uses a compact format for the 3D image representation, composed basically of two images, one color image for the reference view and other grayscale image with the disparity information available for each pixel. Solutions to this problem benefit applications such as Free Viewpoint Television. The biggest challenge is filling in regions without projection information considering the new viewpoint, usually called holes, and other artifacts such as cracks and ghosts that occur due to occlusions and errors in the disparity map. In this dissertation we present techniques for removal and treatment of each of these classes of potential artifacts. The set of proposed methods shows improved results when compared to the current state of the art generation of synthetic views using the DIBR model applied to the Middlebury dataset, considering the SSIM and PSNR metrics.
Advisors/Committee Members: Walter, Marcelo.
Subjects/Keywords: DIBR; Processamento : Imagem; Computacao grafica : Aplicacoes; Hole filling; View synthesis; FTV; 3DTV
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Oliveira, A. Q. d. (2016). Síntese de vistas em depht-image-based rendering (DIBR). (Thesis). Universidade do Rio Grande do Sul. Retrieved from http://hdl.handle.net/10183/143155
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Oliveira, Adriano Quilião de. “Síntese de vistas em depht-image-based rendering (DIBR).” 2016. Thesis, Universidade do Rio Grande do Sul. Accessed December 14, 2019.
http://hdl.handle.net/10183/143155.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Oliveira, Adriano Quilião de. “Síntese de vistas em depht-image-based rendering (DIBR).” 2016. Web. 14 Dec 2019.
Vancouver:
Oliveira AQd. Síntese de vistas em depht-image-based rendering (DIBR). [Internet] [Thesis]. Universidade do Rio Grande do Sul; 2016. [cited 2019 Dec 14].
Available from: http://hdl.handle.net/10183/143155.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Oliveira AQd. Síntese de vistas em depht-image-based rendering (DIBR). [Thesis]. Universidade do Rio Grande do Sul; 2016. Available from: http://hdl.handle.net/10183/143155
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

EPFL
8.
De Abreu Goes, Ana Karina.
Optimized Coding Strategies for Interactive Multiview Video.
Degree: 2015, EPFL
URL: http://infoscience.epfl.ch/record/213655
► The natural next step in improving the realistic experience in multimedia services is interactive multiview video (IMV). IMV promises to enable the users to freely…
(more)
▼ The natural next step in improving the realistic experience in multimedia services is interactive multiview video (IMV). IMV promises to enable the users to freely navigate through a scene by selecting their preferred viewpoints from any
view position for which the corresponding
view is generated. A smooth navigation could be achieved with camera views and views synthesized at the decoder. However, the large amount of data required for such navigation experience still represents a challenge for the current systems, which implies the need for new efficient coding strategies that permit to save on storage and transmission resources, while preserving interactivity in the navigation. In this thesis, we focus on the optimization of coding strategies for IMV systems. First, we address the issues related to the coding techniques for IMV in a multiview video plus depth (MVD) scenario, where texture and depth maps are available for
view synthesis at the decoder. We propose a low complexity algorithm for the selection of the interview prediction structures (PSs) and associated texture and depth quantization parameters (QPs) for IMV under transmission and storage constraints. Simulation results show that our novel low complexity algorithm has near-optimal compression efficiency while preserving interactivity properties at the decoder. Then, considering the limited and heterogeneous capabilities of current networks and decoding devices, we propose a novel adaptive solution for IMV based on a layered multiview representation where camera views are organized into layered subsets to offer different levels of navigation quality depending on the different client constraints. We propose an optimal and a reduced computational complexity greedy algorithms that jointly select the different
view subsets and their encoding rates. Simulation results show the good performance of our novel algorithms compared to a baseline algorithm, proving that an effective IMV adaptive solution should consider the scene content, the client capabilities and their preferences, in building adaptive systems for multiview navigation. Finally, we build on the solution proposed in our second problem and present a general solution to rate allocation problems in multiview video. In particular, we propose a new algorithm to find the optimal Lagrange multiplier in a Lagrangian-based rate allocation problem. We show the performance of our proposed algorithm in both multiview and monoview video scenarios and show that the proposed method is able to compete with complex state-of-the-art rate control techniques. In summary, this thesis addresses important issues for coding multiview video in the design of efficient IMV systems under resource constraints. Our algorithm to select the optimal PS and QPs in a MVD scenario can improve the quality of the rendered views and it can indeed provide new insights for a deeper understanding of specific IMV coding requirements. We show that our algorithm for a layered representation of multiview video provides an effective…
Advisors/Committee Members: Frossard, Pascal, Bernardo Pereira, Fernando Manuel.
Subjects/Keywords: Interactive multiview video (IMV); multiview video plus depth (MVD); navigation; streaming; view synthesis
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
De Abreu Goes, A. K. (2015). Optimized Coding Strategies for Interactive Multiview Video. (Thesis). EPFL. Retrieved from http://infoscience.epfl.ch/record/213655
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
De Abreu Goes, Ana Karina. “Optimized Coding Strategies for Interactive Multiview Video.” 2015. Thesis, EPFL. Accessed December 14, 2019.
http://infoscience.epfl.ch/record/213655.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
De Abreu Goes, Ana Karina. “Optimized Coding Strategies for Interactive Multiview Video.” 2015. Web. 14 Dec 2019.
Vancouver:
De Abreu Goes AK. Optimized Coding Strategies for Interactive Multiview Video. [Internet] [Thesis]. EPFL; 2015. [cited 2019 Dec 14].
Available from: http://infoscience.epfl.ch/record/213655.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
De Abreu Goes AK. Optimized Coding Strategies for Interactive Multiview Video. [Thesis]. EPFL; 2015. Available from: http://infoscience.epfl.ch/record/213655
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Université de Lorraine
9.
Rolin, Pierre.
Synthèse de vues pour l’initialisation de pose : Viewpoint synthesis for pose initialisation.
Degree: Docteur es, Informatique, 2017, Université de Lorraine
URL: http://www.theses.fr/2017LORR0025
► La localisation est un problème récurrent de la vision par ordinateur, avec des applications dans des domaines multiples tels que la robotique ou la réalité…
(more)
▼ La localisation est un problème récurrent de la vision par ordinateur, avec des applications dans des domaines multiples tels que la robotique ou la réalité augmentée. Dans cette thèse on considère en particulier le problème d'initialisation de la pose, c'est-à-dire la localisation sans information a priori sur la position de la caméra. Nous nous intéressons à la localisation à partir d'une image monoculaire et d'un nuage de points reconstruit à partir d'une séquence d'images. Puisque nous n'avons pas d'a priori sur la position de la caméra, l'estimation de la pose s'appuie sur la recherche de correspondances entre des points de l'image et des points du modèle de la scène. Cette mise en correspondance est difficile en raison de sa combinatoire élevée. Elle peut être mise en défaut lorsque l'image dont on cherche la pose est très différente de celles ayant servi à la construction du modèle, en particulier en présence de forts changements de point de vue. Cette thèse développe une approche permettant la mise en correspondance image-modèle dans ces situations complexes. Elle consiste à synthétiser localement l'apparence de la scène à partir de points de vue virtuels puis à ajouter au modèle des descripteurs extraits des images synthétisées. Comme le modèle de scène est un nuage de points, la synthèse n'est pas faite par rendu 3D mais utilise des transformations 2D locales des observations connues de la scène. Les contributions suivantes sont apportées. Nous étudions différents modèles de transformation possibles et montrons que la synthèse par homographie est la plus adaptée pour ce type d'application. Nous définissons une méthode de positionnement des points de vue virtuels par rapport à une segmentation de la scène en patchs plans. Nous assurons l'efficacité de l'approche proposée en ne synthétisant que des vues utiles : elles sont éloignées de celles existantes et elles ne se recouvrent pas. Nous vérifions également que la scène est visible à partir des points des vue virtuels pour ne pas produire des vues aberrantes à cause d’occultations. Enfin, nous proposons une méthode de recherche de correspondances image-modèle qui est à la fois rapide et robuste. Cette méthode exploite la répartition non-uniforme des correspondances correctes dans le modèle, ce qui permet de guider leur recherche. Les résultats expérimentaux montrent que la méthode proposée permet de calculer des poses dans des configurations défavorables où les approches standard échouent. De façon générale la précision des poses obtenues augmente significativement lorsque la synthèse de vue est utilisée. Enfin nous montrons que, en facilitant la mise en correspondance image-modèle, cette méthode accélère le calcul de pose
Localisation is a central problem of computer vision which has numerous applications such as robotics or augmented reality. In this thesis we consider the problem of pose initialisation, which is pose computation without prior knowledge on the camera position. We are interested in pose computation from a single image and a point cloud…
Advisors/Committee Members: Berger, Marie-Odile (thesis director), Sur, Frédéric (thesis director).
Subjects/Keywords: Calcul de pose; Mise en correspondance; Synthèse de vues; Pose computation; Matching; View synthesis; 006.37
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Rolin, P. (2017). Synthèse de vues pour l’initialisation de pose : Viewpoint synthesis for pose initialisation. (Doctoral Dissertation). Université de Lorraine. Retrieved from http://www.theses.fr/2017LORR0025
Chicago Manual of Style (16th Edition):
Rolin, Pierre. “Synthèse de vues pour l’initialisation de pose : Viewpoint synthesis for pose initialisation.” 2017. Doctoral Dissertation, Université de Lorraine. Accessed December 14, 2019.
http://www.theses.fr/2017LORR0025.
MLA Handbook (7th Edition):
Rolin, Pierre. “Synthèse de vues pour l’initialisation de pose : Viewpoint synthesis for pose initialisation.” 2017. Web. 14 Dec 2019.
Vancouver:
Rolin P. Synthèse de vues pour l’initialisation de pose : Viewpoint synthesis for pose initialisation. [Internet] [Doctoral dissertation]. Université de Lorraine; 2017. [cited 2019 Dec 14].
Available from: http://www.theses.fr/2017LORR0025.
Council of Science Editors:
Rolin P. Synthèse de vues pour l’initialisation de pose : Viewpoint synthesis for pose initialisation. [Doctoral Dissertation]. Université de Lorraine; 2017. Available from: http://www.theses.fr/2017LORR0025

University of Kentucky
10.
Guinnip, David.
Master Texture Space: An Efficient Encoding for Projectively Mapped Objects.
Degree: 2005, University of Kentucky
URL: http://uknowledge.uky.edu/gradschool_theses/228
► Projectively textured models are used in an increasingly large number of applicationsthat dynamically combine images with a simple geometric surface in a viewpoint dependentway. These…
(more)
▼ Projectively textured models are used in an increasingly large number of applicationsthat dynamically combine images with a simple geometric surface in a viewpoint dependentway. These models can provide visual fidelity while retaining the effects affordedby geometric approximation such as shadow casting and accurate perspective distortion.However, the number of stored views can be quite large and novel views must be synthesizedduring the rendering process because no single view may correctly texture the entireobject surface. This work introduces the Master Texture encoding and demonstrates thatthe encoding increases the utility of projectively textured objects by reducing render-timeoperations. Encoding involves three steps; 1) all image regions that correspond to the samegeometric mesh element are extracted and warped to a facet of uniform size and shape,2) an efficient packing of these facets into a new Master Texture image is computed, and3) the visibility of each pixel in the new Master Texture data is guaranteed using a simplealgorithm to discard occluded pixels in each view. Because the encoding implicitly representsthe multi-view geometry of the multiple images, a single texture mesh is sufficientto render the view-dependent model. More importantly, every Master Texture image cancorrectly texture the entire surface of the object, removing expensive computations suchas visibility analysis from the rendering algorithm. A benefit of this encoding is the supportfor pixel-wise view synthesis. The utility of pixel-wise view synthesis is demonstratedwith a real-time Master Texture encoded VDTM application. Pixel-wise synthesis is alsodemonstrated with an algorithm that distills a set of Master Texture images to a singleview-independent Master Texture image.
Subjects/Keywords: View-dependent Projective Texture Mapping; Visiblity Analysis; View Synthesis; Triangle Packing; Rectangle Packing
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Guinnip, D. (2005). Master Texture Space: An Efficient Encoding for Projectively Mapped Objects. (Masters Thesis). University of Kentucky. Retrieved from http://uknowledge.uky.edu/gradschool_theses/228
Chicago Manual of Style (16th Edition):
Guinnip, David. “Master Texture Space: An Efficient Encoding for Projectively Mapped Objects.” 2005. Masters Thesis, University of Kentucky. Accessed December 14, 2019.
http://uknowledge.uky.edu/gradschool_theses/228.
MLA Handbook (7th Edition):
Guinnip, David. “Master Texture Space: An Efficient Encoding for Projectively Mapped Objects.” 2005. Web. 14 Dec 2019.
Vancouver:
Guinnip D. Master Texture Space: An Efficient Encoding for Projectively Mapped Objects. [Internet] [Masters thesis]. University of Kentucky; 2005. [cited 2019 Dec 14].
Available from: http://uknowledge.uky.edu/gradschool_theses/228.
Council of Science Editors:
Guinnip D. Master Texture Space: An Efficient Encoding for Projectively Mapped Objects. [Masters Thesis]. University of Kentucky; 2005. Available from: http://uknowledge.uky.edu/gradschool_theses/228

University of Southern California
11.
Kim, Woo-Shik.
3-D video coding system with enhanced rendered view
quality.
Degree: PhD, Electrical Engineering, 2011, University of Southern California
URL: http://digitallibrary.usc.edu/cdm/compoundobject/collection/p15799coll127/id/636333/rec/17
► The objective of this research is to develop a new 3-D video coding system which can provide better coding efficiency with improved subjective quality as…
(more)
▼ The objective of this research is to develop a new 3-D
video coding system which can provide better coding efficiency with
improved subjective quality as compared to existing 3-D video
systems such as the depth image based rendering (DIBR) system.
Clearly, one would be able to increase overall performance by
focusing on better “generic” coding tools. Instead, here we focus
on techniques that are specific of 3-D video. Specifically, we
consider improved representations for depth information as well as
information that can directly contribute to improved intermediate
view interpolation. ❧ As a starting point, we analyze the
distortions that occur in rendered views generated using the DIBR
system, and classify them in order to evaluate their impact on
subjective quality. As a result, we find that the rendered
view
distortion due to depth map coding has non-linear characteristics
(i.e., increases in intensity errors in the interpolated
view are
not proportional to increases in depth map coding errors) and is
highly localized (i.e., very large errors occur only in a small
subset of pixels in a video frame), which can lead to significant
degradation in perceptual quality. A flickering artifact is also
observed due to temporal variation of depth map sequence. ❧ To
solve these problems, we first propose new coding tools which can
reduce the rendered
view distortion by defining a new distortion
metric to derive relationships between distortions in coded depth
map and rendered
view. In addition, a new skip mode selection
method is proposed based on local video characteristics. Our
experimental results show the efficiency of the proposed method
with coding gains of up to 1.6 dB in interpolated frame quality as
well as better subjective quality with reduced flickering
artifacts. ❧ We also propose a new transform coding using graph
based representation of a signal, which we name as graph based
transform. Considering depth map consists of smooth regions with
sharp edges along object boundaries, efficient transform coding can
be performed by forming a graph in which the pixels are not
connected across edges. Experimental results reveal that coding
efficiency improvement of 0.4 dB can be achieved by applying the
new transform in a hybrid manner with DCT to compress a depth map.
❧ Secondly, we propose a solution in which depth transition data is
encoded and transmitted to the decoder. Depth transition data for a
given pixel indicates the camera position for which this pixel’s
depth will change. For example in a pixel corresponding to
foreground in the left image, and background in the right image,
this information helps us determine in which intermediate
view (as
we move left to right), this pixel will become a background pixel.
The main reason to consider transmitting explicitly this
information is that it can be used to improve
view interpolation at
many different intermediate camera positions. Simulation results
show that the subjective quality can be significantly improved
using our proposed depth transition data. Maximum PSNR…
Advisors/Committee Members: Ortega, Antonio (Committee Chair), Kuo, C.-C. Jay (Committee Member), Neumann, Ulrich (Committee Member).
Subjects/Keywords: signal processing; multimedia processing; image processing; video processing; 3-D video; image compression; video compression; video coding; view synthesis; view rendering; depth map coding
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Kim, W. (2011). 3-D video coding system with enhanced rendered view
quality. (Doctoral Dissertation). University of Southern California. Retrieved from http://digitallibrary.usc.edu/cdm/compoundobject/collection/p15799coll127/id/636333/rec/17
Chicago Manual of Style (16th Edition):
Kim, Woo-Shik. “3-D video coding system with enhanced rendered view
quality.” 2011. Doctoral Dissertation, University of Southern California. Accessed December 14, 2019.
http://digitallibrary.usc.edu/cdm/compoundobject/collection/p15799coll127/id/636333/rec/17.
MLA Handbook (7th Edition):
Kim, Woo-Shik. “3-D video coding system with enhanced rendered view
quality.” 2011. Web. 14 Dec 2019.
Vancouver:
Kim W. 3-D video coding system with enhanced rendered view
quality. [Internet] [Doctoral dissertation]. University of Southern California; 2011. [cited 2019 Dec 14].
Available from: http://digitallibrary.usc.edu/cdm/compoundobject/collection/p15799coll127/id/636333/rec/17.
Council of Science Editors:
Kim W. 3-D video coding system with enhanced rendered view
quality. [Doctoral Dissertation]. University of Southern California; 2011. Available from: http://digitallibrary.usc.edu/cdm/compoundobject/collection/p15799coll127/id/636333/rec/17
12.
Meilland, Maxime.
Cartographie RGB-D dense pour la localisation visuelle temps-réel et la navigation autonome : Dense RGB-D mapping for real-time localisation and autonomous navigation.
Degree: Docteur es, Informatique temps réel, robotique et automatique, 2012, Paris, ENMP
URL: http://www.theses.fr/2007ENMP0007
► Dans le contexte de la navigation autonome en environnement urbain, une localisation précise du véhicule est importante pour une navigation sure et fiable. La faible…
(more)
▼ Dans le contexte de la navigation autonome en environnement urbain, une localisation précise du véhicule est importante pour une navigation sure et fiable. La faible précision des capteurs bas coût existants tels que le système GPS, nécessite l'utilisation d'autres capteurs eux aussi à faible coût. Les caméras mesurent une information photométrique riche et précise sur l'environnement, mais nécessitent l'utilisation d'algorithmes de traitement avancés pour obtenir une information sur la géométrie et sur la position de la caméra dans l'environnement. Cette problématique est connue sous le terme de Cartographie et Localisation Simultanées (SLAM visuel). En général, les techniques de SLAM sont incrémentales et dérivent sur de longues trajectoires. Pour simplifier l'étape de localisation, il est proposé de découpler la partie cartographie et la partie localisation en deux phases: la carte est construite hors-ligne lors d'une phase d'apprentissage, et la localisation est effectuée efficacement en ligne à partir de la carte 3D de l'environnement. Contrairement aux approches classiques, qui utilisent un modèle 3D global approximatif, une nouvelle représentation égo-centrée dense est proposée. Cette représentation est composée d'un graphe d'images sphériques augmentées par l'information dense de profondeur (RGB+D), et permet de cartographier de larges environnements. Lors de la localisation en ligne, ce type de modèle apporte toute l'information nécessaire pour une localisation précise dans le voisinage du graphe, et permet de recaler en temps-réel l'image perçue par une caméra embarquée sur un véhicule, avec les images du graphe, en utilisant une technique d'alignement d'images directe. La méthode de localisation proposée, est précise, robuste aux aberrations et prend en compte les changements d'illumination entre le modèle de la base de données et les images perçues par la caméra. Finalement, la précision et la robustesse de la localisation permettent à un véhicule autonome, équipé d'une caméra, de naviguer de façon sure en environnement urbain.
In an autonomous navigation context, a precise localisation of the vehicule is important to ensure a reliable navigation. Low cost sensors such as GPS systems are inacurrate and inefficicent in urban areas, and therefore the employ of such sensors alone is not well suited for autonomous navigation. On the other hand, camera sensors provide a dense photometric measure that can be processed to obtain both localisation and mapping information. In the robotics community, this problem is well known as Simultaneous Localisation and Mapping (SLAM) and it has been studied for the last thirty years. In general, SLAM algorithms are incremental and prone to drift, thus such methods may not be efficient in large scale environments for real-time localisation. Clearly, an a-priori 3D model simplifies the localisation and navigation tasks since it allows to decouple the structure and motion estimation problems. Indeed, the map can be previously computed during a learning phase, whilst the…
Advisors/Committee Members: Rives, Patrick (thesis director), Comport, Andrew (thesis director).
Subjects/Keywords: SLAM; Navigation; Localisation; Suivi visuel; Synthèse de nouvelle vue; Cartographie; SLAM; Navigation; Localisation; Visual Tracking; Novel View Synthesis; Mapping
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Meilland, M. (2012). Cartographie RGB-D dense pour la localisation visuelle temps-réel et la navigation autonome : Dense RGB-D mapping for real-time localisation and autonomous navigation. (Doctoral Dissertation). Paris, ENMP. Retrieved from http://www.theses.fr/2007ENMP0007
Chicago Manual of Style (16th Edition):
Meilland, Maxime. “Cartographie RGB-D dense pour la localisation visuelle temps-réel et la navigation autonome : Dense RGB-D mapping for real-time localisation and autonomous navigation.” 2012. Doctoral Dissertation, Paris, ENMP. Accessed December 14, 2019.
http://www.theses.fr/2007ENMP0007.
MLA Handbook (7th Edition):
Meilland, Maxime. “Cartographie RGB-D dense pour la localisation visuelle temps-réel et la navigation autonome : Dense RGB-D mapping for real-time localisation and autonomous navigation.” 2012. Web. 14 Dec 2019.
Vancouver:
Meilland M. Cartographie RGB-D dense pour la localisation visuelle temps-réel et la navigation autonome : Dense RGB-D mapping for real-time localisation and autonomous navigation. [Internet] [Doctoral dissertation]. Paris, ENMP; 2012. [cited 2019 Dec 14].
Available from: http://www.theses.fr/2007ENMP0007.
Council of Science Editors:
Meilland M. Cartographie RGB-D dense pour la localisation visuelle temps-réel et la navigation autonome : Dense RGB-D mapping for real-time localisation and autonomous navigation. [Doctoral Dissertation]. Paris, ENMP; 2012. Available from: http://www.theses.fr/2007ENMP0007
13.
Meilland, Maxime.
Cartographie RGB-D dense pour la localisation visuelle temps-réel et la navigation autonome : Dense RGB-D mapping for real-time localisation and autonomous navigation.
Degree: Docteur es, Informatique temps réel, robotique et automatique, 2012, Paris, ENMP
URL: http://www.theses.fr/2012ENMP0007
► Dans le contexte de la navigation autonome en environnement urbain, une localisation précise du véhicule est importante pour une navigation sure et fiable. La faible…
(more)
▼ Dans le contexte de la navigation autonome en environnement urbain, une localisation précise du véhicule est importante pour une navigation sure et fiable. La faible précision des capteurs bas coût existants tels que le système GPS, nécessite l'utilisation d'autres capteurs eux aussi à faible coût. Les caméras mesurent une information photométrique riche et précise sur l'environnement, mais nécessitent l'utilisation d'algorithmes de traitement avancés pour obtenir une information sur la géométrie et sur la position de la caméra dans l'environnement. Cette problématique est connue sous le terme de Cartographie et Localisation Simultanées (SLAM visuel). En général, les techniques de SLAM sont incrémentales et dérivent sur de longues trajectoires. Pour simplifier l'étape de localisation, il est proposé de découpler la partie cartographie et la partie localisation en deux phases: la carte est construite hors-ligne lors d'une phase d'apprentissage, et la localisation est effectuée efficacement en ligne à partir de la carte 3D de l'environnement. Contrairement aux approches classiques, qui utilisent un modèle 3D global approximatif, une nouvelle représentation égo-centrée dense est proposée. Cette représentation est composée d'un graphe d'images sphériques augmentées par l'information dense de profondeur (RGB+D), et permet de cartographier de larges environnements. Lors de la localisation en ligne, ce type de modèle apporte toute l'information nécessaire pour une localisation précise dans le voisinage du graphe, et permet de recaler en temps-réel l'image perçue par une caméra embarquée sur un véhicule, avec les images du graphe, en utilisant une technique d'alignement d'images directe. La méthode de localisation proposée, est précise, robuste aux aberrations et prend en compte les changements d'illumination entre le modèle de la base de données et les images perçues par la caméra. Finalement, la précision et la robustesse de la localisation permettent à un véhicule autonome, équipé d'une caméra, de naviguer de façon sure en environnement urbain.
In an autonomous navigation context, a precise localisation of the vehicule is important to ensure a reliable navigation. Low cost sensors such as GPS systems are inacurrate and inefficicent in urban areas, and therefore the employ of such sensors alone is not well suited for autonomous navigation. On the other hand, camera sensors provide a dense photometric measure that can be processed to obtain both localisation and mapping information. In the robotics community, this problem is well known as Simultaneous Localisation and Mapping (SLAM) and it has been studied for the last thirty years. In general, SLAM algorithms are incremental and prone to drift, thus such methods may not be efficient in large scale environments for real-time localisation. Clearly, an a-priori 3D model simplifies the localisation and navigation tasks since it allows to decouple the structure and motion estimation problems. Indeed, the map can be previously computed during a learning phase, whilst the…
Advisors/Committee Members: Rives, Patrick (thesis director), Comport, Andrew Ian (thesis director).
Subjects/Keywords: SLAM; Navigation; Localisation; Suivi visuel; Synthèse de nouvelle vue; Cartographie; SLAM; Navigation; Localisation; Visual Tracking; Novel View Synthesis; Mapping
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Meilland, M. (2012). Cartographie RGB-D dense pour la localisation visuelle temps-réel et la navigation autonome : Dense RGB-D mapping for real-time localisation and autonomous navigation. (Doctoral Dissertation). Paris, ENMP. Retrieved from http://www.theses.fr/2012ENMP0007
Chicago Manual of Style (16th Edition):
Meilland, Maxime. “Cartographie RGB-D dense pour la localisation visuelle temps-réel et la navigation autonome : Dense RGB-D mapping for real-time localisation and autonomous navigation.” 2012. Doctoral Dissertation, Paris, ENMP. Accessed December 14, 2019.
http://www.theses.fr/2012ENMP0007.
MLA Handbook (7th Edition):
Meilland, Maxime. “Cartographie RGB-D dense pour la localisation visuelle temps-réel et la navigation autonome : Dense RGB-D mapping for real-time localisation and autonomous navigation.” 2012. Web. 14 Dec 2019.
Vancouver:
Meilland M. Cartographie RGB-D dense pour la localisation visuelle temps-réel et la navigation autonome : Dense RGB-D mapping for real-time localisation and autonomous navigation. [Internet] [Doctoral dissertation]. Paris, ENMP; 2012. [cited 2019 Dec 14].
Available from: http://www.theses.fr/2012ENMP0007.
Council of Science Editors:
Meilland M. Cartographie RGB-D dense pour la localisation visuelle temps-réel et la navigation autonome : Dense RGB-D mapping for real-time localisation and autonomous navigation. [Doctoral Dissertation]. Paris, ENMP; 2012. Available from: http://www.theses.fr/2012ENMP0007

McMaster University
14.
Rezaee Kaviani, Hoda.
Novel Image Interpolation Schemes with Applications to Frame Rate Conversion and View Synthesis.
Degree: PhD, 2018, McMaster University
URL: http://hdl.handle.net/11375/24296
► Image interpolation is the process of generating a new image utilizing a set of available images. The available images may be taken with a camera…
(more)
▼ Image interpolation is the process of generating a new image utilizing a set of available images. The available images may be taken with a camera at different times, or with multiple cameras and from different viewpoints. Usually, the interpolation problem in the first scenario is called Frame Rate-Up Conversion (FRUC), and the second one
view synthesis.
This thesis focuses on image interpolation and addresses both FRUC and
view synthesis problems. We propose a novel FRUC method using optical flow motion estimation and a patch-based reconstruction scheme. FRUC interpolates new frames between original frames of a video to increase the number of frames, and increases motion continuity.
In our approach first, forward and backward motion vectors are obtained using an optical flow algorithm, and reconstructed versions of the current and previous frames are generated by our patch-based reconstruction scheme.
Using the original and reconstructed versions of the current and previous frames, two mismatch masks are obtained. Then two versions of the middle frame are generated using a patch-based scheme, with estimated motion vectors and the current and previous frames. Finally, a middle mask, which identifies the mismatch areas of the two middle frames is reconstructed. Using these three masks, the best candidates for interpolation are selected and fused to obtain the final middle frame.
Due to the patch-based nature of our interpolation scheme most of the holes and cracks will be filled.
Although there is always a probability of having holes, the size and number of such holes are much smaller than those that would be generated using pixel-based mapping. The rare holes are filled using existing hole-filling algorithms. With fewer and smaller holes, simpler hole-filling algorithms can be applied to the image and the overall complexity of the required post processing decreases.
View synthesis is the process of generating a new (virtual)
view using available ones. Depending on the amount of available geometric information,
view synthesis techniques can be divided into three categories: Image Based Rendering (IBR), Depth Image Based Rendering (DIBR), and Model Based Rendering (MBR).
We introduce an adaptive version, patch-based scheme for IBR. This patch-based scheme reduces the size and number of holes during reconstruction. The size of patch is determined in response to edge information for better reconstruction, especially near the boundaries. In the first stage of the algorithm, disparity is obtained using optical flow estimation. Then, a reconstructed version of the left and right views are generated using our adaptive patch-based algorithm. The mismatches between each
view and its reconstructed version are obtained in the mismatch detection steps.
This stage results in two masks as outputs, which help with the refinement of disparities and the selection of the best patches for final
synthesis. Finally, the remaining holes are filled using our simple hole filling scheme and the refined disparities. The…
Advisors/Committee Members: Shirani, Shahram, Electrical and Computer Engineering.
Subjects/Keywords: Image interpolation; View synthesis; Frame rate conversion; Image processing; Depth image based rendering; Free viewpoint TV
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Rezaee Kaviani, H. (2018). Novel Image Interpolation Schemes with Applications to Frame Rate Conversion and View Synthesis. (Doctoral Dissertation). McMaster University. Retrieved from http://hdl.handle.net/11375/24296
Chicago Manual of Style (16th Edition):
Rezaee Kaviani, Hoda. “Novel Image Interpolation Schemes with Applications to Frame Rate Conversion and View Synthesis.” 2018. Doctoral Dissertation, McMaster University. Accessed December 14, 2019.
http://hdl.handle.net/11375/24296.
MLA Handbook (7th Edition):
Rezaee Kaviani, Hoda. “Novel Image Interpolation Schemes with Applications to Frame Rate Conversion and View Synthesis.” 2018. Web. 14 Dec 2019.
Vancouver:
Rezaee Kaviani H. Novel Image Interpolation Schemes with Applications to Frame Rate Conversion and View Synthesis. [Internet] [Doctoral dissertation]. McMaster University; 2018. [cited 2019 Dec 14].
Available from: http://hdl.handle.net/11375/24296.
Council of Science Editors:
Rezaee Kaviani H. Novel Image Interpolation Schemes with Applications to Frame Rate Conversion and View Synthesis. [Doctoral Dissertation]. McMaster University; 2018. Available from: http://hdl.handle.net/11375/24296
15.
Tian, Shishun.
Image Quality Assessment of 3D Synthesized Views : Évaluation de la qualité des images obtenues par synthèse de vues 3D.
Degree: Docteur es, Signal, Image, Vision, 2019, Rennes, INSA
URL: http://www.theses.fr/2019ISAR0002
► Depth-Image-Based Rendering (DIBR) est une technologie fondamentale dans plusieurs applications liées à la 3D, telles que la vidéo en mode point de vue libre (FVV),…
(more)
▼ Depth-Image-Based Rendering (DIBR) est une technologie fondamentale dans plusieurs applications liées à la 3D, telles que la vidéo en mode point de vue libre (FVV), la réalité virtuelle (VR) et la réalité augmentée (AR). Cependant, l'évaluation de la qualité des vues synthétisées par DIBR a également posé de nouveaux problèmes, car ce processus induit de nouveaux types de distorsions, qui sont intrinsèquement différentes des distorsions provoquées par le codage vidéo. Ce travail est destiné à mieux évaluer la qualité des vues synthétisées par DIBR en multimédia immersif. Au chapitre 2, nous proposons deux métriques complètements sans référence (NR). Le principe de la première métrique NR NIQSV consiste à utiliser plusieurs opérations morphologiques d’ouverture et de fermeture pour détecter et mesurer les distorsions, telles que les régions floues et l’effritement. Dans la deuxième métrique NR NIQSV+, nous améliorons NIQSV en ajoutant un détecteur de “black hole” et une détection “stretching”.Au chapitre 3, nous proposons deux métriques de référence complète pour traiter les distorsions géométriques à l'aide d'un masque de désocclusion et d'une méthode de correspondance de blocs multi-résolution. Au chapitre 4, nous présentons une nouvelle base de données d'images synthétisée par DIBR avec ses scores subjectifs associés. Ce travail se concentre sur les distorsions uniquement induites par différentes méthodes de synthèse de DIBR qui déterminent la qualité d’expérience (QoE) de ces applications liées à DIBR. En outre, nous effectuons également une analyse de référence des mesures d'évaluation de la qualité objective de pointe pour les vues synthétisées par DIBR sur cette base de données. Le chapitre 5 conclut les contributions de cette thèse et donne quelques orientations pour les travaux futurs.
Depth-Image-Based Rendering (DIBR) is a fundamental technology in several 3D-related applications, such as Free viewpoint video (FVV), Virtual Reality (VR) and Augmented Reality (AR). However, new challenges have also been brought in assessing the quality of DIBR-synthesized views since this process induces some new types of distortions, which are inherently different from the distortions caused by video coding. This work is dedicated to better evaluate the quality of DIBRsynthesized views in immersive multimedia. In chapter 2, we propose a completely No-reference (NR) metric. The principle of the first NR metrics NIQSV is to use a couple of opening and closing morphological operations to detect and measure the distortions, such as “blurry regions” and “crumbling”. In the second NR metric NIQSV+, we improve NIQSV by adding a “black hole” and a “stretching” detection. In chapter 3, we propose two Fullreference metrics to handle the geometric distortions by using a dis-occlusion mask and a multi-resolution block matching methods.In chapter 4, we present a new DIBR-synthesized image database with its associated subjective scores. This work focuses on the distortions only induced by different DIBR synthesis methods which…
Advisors/Committee Members: Morin, Luce (thesis director), Zhang, Lu (thesis director).
Subjects/Keywords: Depth-image-based-rendering (DIBR); View synthesis; Distortion; Image quality assessment; Quality of Experience (QoE); : Depth-image-based-rendering (DIBR); View synthesis; Distortion; Image quality assessment; Quality of Experience (QoE); 621.388
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Tian, S. (2019). Image Quality Assessment of 3D Synthesized Views : Évaluation de la qualité des images obtenues par synthèse de vues 3D. (Doctoral Dissertation). Rennes, INSA. Retrieved from http://www.theses.fr/2019ISAR0002
Chicago Manual of Style (16th Edition):
Tian, Shishun. “Image Quality Assessment of 3D Synthesized Views : Évaluation de la qualité des images obtenues par synthèse de vues 3D.” 2019. Doctoral Dissertation, Rennes, INSA. Accessed December 14, 2019.
http://www.theses.fr/2019ISAR0002.
MLA Handbook (7th Edition):
Tian, Shishun. “Image Quality Assessment of 3D Synthesized Views : Évaluation de la qualité des images obtenues par synthèse de vues 3D.” 2019. Web. 14 Dec 2019.
Vancouver:
Tian S. Image Quality Assessment of 3D Synthesized Views : Évaluation de la qualité des images obtenues par synthèse de vues 3D. [Internet] [Doctoral dissertation]. Rennes, INSA; 2019. [cited 2019 Dec 14].
Available from: http://www.theses.fr/2019ISAR0002.
Council of Science Editors:
Tian S. Image Quality Assessment of 3D Synthesized Views : Évaluation de la qualité des images obtenues par synthèse de vues 3D. [Doctoral Dissertation]. Rennes, INSA; 2019. Available from: http://www.theses.fr/2019ISAR0002

University of Central Florida
16.
Berkowitz, Phillip.
A Statistical Approach To View Synthesis.
Degree: 2009, University of Central Florida
URL: https://stars.library.ucf.edu/etd/4173
► View Synthesis is the challenging problem of predicting a new view or pose of an object given an exemplar view or set of views. This…
(more)
▼ View Synthesis is the challenging problem of predicting a new
view or pose of an object given an exemplar
view or set of views. This thesis presents a novel approach for the problem of
view synthesis. The proposed method uses global features rather than local geometry to achieve an effect similar to that of the well known
view morphing method . While previous approaches to the
view synthesis problem have shown impressive results, they are highly dependent on being able to solve for epipolar geometry and therefore have a very precise correspondence between reference images. In cases where this is not possible such as noisy data, low contrast data, or long wave infrared data an alternative approach is desirable. Here two problems will be considered. The proposed
view synthesis method will be used to
synthesis new views given a set of reference views. Additionally the algorithm will be extended to
synthesis new lighting conditions and thermal signatures. Finally the algorithm will be applied toward enhancing the ATR problem by creating additional training data to increase the likelihood of detection and classification.
Advisors/Committee Members: Shah, Mubarak.
Subjects/Keywords: View Synthesis; View Morphing; ATR; EOIR; Electrical and Computer Engineering; Electrical and Electronics; Engineering
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Berkowitz, P. (2009). A Statistical Approach To View Synthesis. (Masters Thesis). University of Central Florida. Retrieved from https://stars.library.ucf.edu/etd/4173
Chicago Manual of Style (16th Edition):
Berkowitz, Phillip. “A Statistical Approach To View Synthesis.” 2009. Masters Thesis, University of Central Florida. Accessed December 14, 2019.
https://stars.library.ucf.edu/etd/4173.
MLA Handbook (7th Edition):
Berkowitz, Phillip. “A Statistical Approach To View Synthesis.” 2009. Web. 14 Dec 2019.
Vancouver:
Berkowitz P. A Statistical Approach To View Synthesis. [Internet] [Masters thesis]. University of Central Florida; 2009. [cited 2019 Dec 14].
Available from: https://stars.library.ucf.edu/etd/4173.
Council of Science Editors:
Berkowitz P. A Statistical Approach To View Synthesis. [Masters Thesis]. University of Central Florida; 2009. Available from: https://stars.library.ucf.edu/etd/4173

University of Central Florida
17.
Xiao, Jiangjian.
Image Based View Synthesis.
Degree: 2004, University of Central Florida
URL: https://stars.library.ucf.edu/etd/263
► This dissertation deals with the image-based approach to synthesize a virtual scene using sparse images or a video sequence without the use of 3D models.…
(more)
▼ This dissertation deals with the image-based approach to synthesize a virtual scene using sparse images or a video sequence without the use of 3D models. In our scenario, a real dynamic or static scene is captured by a set of un-calibrated images from different viewpoints. After automatically recovering the geometric transformations between these images, a series of photo-realistic virtual views can be rendered and a virtual environment covered by these several static cameras can be synthesized. This image-based approach has applications in object recognition, object transfer, video
synthesis and video compression. In this dissertation, I have contributed to several sub-problems related to image based
view synthesis. Before image-based
view synthesis can be performed, images need to be segmented into individual objects. Assuming that a scene can approximately be described by multiple planar regions, I have developed a robust and novel approach to automatically extract a set of affine or projective transformations induced by these regions, correctly detect the occlusion pixels over multiple consecutive frames, and accurately segment the scene into several motion layers. First, a number of seed regions using correspondences in two frames are determined, and the seed regions are expanded and outliers are rejected employing the graph cuts method integrated with level set representation. Next, these initial regions are merged into several initial layers according to the motion similarity. Third, the occlusion order constraints on multiple frames are explored, which guarantee that the occlusion area increases with the temporal order in a short period and effectively maintains segmentation consistency over multiple consecutive frames. Then the correct layer segmentation is obtained by using a graph cuts algorithm, and the occlusions between the overlapping layers are explicitly determined. Several experimental results are demonstrated to show that our approach is effective and robust. Recovering the geometrical transformations among images of a scene is a prerequisite step for image-based
view synthesis. I have developed a wide baseline matching algorithm to identify the correspondences between two un-calibrated images, and to further determine the geometric relationship between images, such as epipolar geometry or projective transformation. In our approach, a set of salient features, edge-corners, are detected to provide robust and consistent matching primitives. Then, based on the Singular Value Decomposition (SVD) of an affine matrix, we effectively quantize the search space into two independent subspaces for rotation angle and scaling factor, and then we use a two-stage affine matching algorithm to obtain robust matches between these two frames. The experimental results on a number of wide baseline images strongly demonstrate that our matching method outperforms the state-of-art algorithms even under the significant camera motion, illumination variation, occlusion, and self-similarity. Given the wide baseline matches…
Advisors/Committee Members: Shah, Mubarak.
Subjects/Keywords: Wide baseline matching; motion segmentation; view morphing; multiple view geometry; video synthesis.; Computer Sciences; Engineering
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Xiao, J. (2004). Image Based View Synthesis. (Doctoral Dissertation). University of Central Florida. Retrieved from https://stars.library.ucf.edu/etd/263
Chicago Manual of Style (16th Edition):
Xiao, Jiangjian. “Image Based View Synthesis.” 2004. Doctoral Dissertation, University of Central Florida. Accessed December 14, 2019.
https://stars.library.ucf.edu/etd/263.
MLA Handbook (7th Edition):
Xiao, Jiangjian. “Image Based View Synthesis.” 2004. Web. 14 Dec 2019.
Vancouver:
Xiao J. Image Based View Synthesis. [Internet] [Doctoral dissertation]. University of Central Florida; 2004. [cited 2019 Dec 14].
Available from: https://stars.library.ucf.edu/etd/263.
Council of Science Editors:
Xiao J. Image Based View Synthesis. [Doctoral Dissertation]. University of Central Florida; 2004. Available from: https://stars.library.ucf.edu/etd/263

University of Ottawa
18.
Brunton, Alan P.
Multi-scale Methods for Omnidirectional Stereo with Application to Real-time Virtual Walkthroughs
.
Degree: 2012, University of Ottawa
URL: http://hdl.handle.net/10393/23552
► This thesis addresses a number of problems in computer vision, image processing, and geometry processing, and presents novel solutions to these problems. The overarching theme…
(more)
▼ This thesis addresses a number of problems in computer vision, image processing, and geometry processing, and presents novel solutions to these problems. The overarching theme of the techniques presented here is a multi-scale approach, leveraging mathematical tools to represent images and surfaces at different scales, and methods that can be adapted from one type of domain (eg., the plane) to another (eg., the sphere). The main problem addressed in this thesis is known as stereo reconstruction: reconstructing the geometry of a scene or object from two or more images of that scene. We develop novel algorithms to do this, which work for both planar and spherical images. By developing a novel way to formulate the notion of disparity for spherical images, we are able effectively adapt our algorithms from planar to spherical images. Our stereo reconstruction algorithm is based on a novel application of distance transforms to multi-scale matching. We use matching information aggregated over multiple scales, and enforce consistency between these scales using distance transforms. We then show how multiple spherical disparity maps can be efficiently and robustly fused using visibility and other geometric constraints. We then show how the reconstructed point clouds can be used to synthesize a realistic sequence of novel views, images from points of view not captured in the input images, in real-time. Along the way to this result, we address some related problems. For example, multi-scale features can be detected in spherical images by convolving those images with a filterbank, generating an overcomplete spherical wavelet representation of the image from which the multiscale features can be extracted. Convolution of spherical images is much more efficient in the spherical harmonic domain than in the spatial domain. Thus, we develop a GPU implementation for fast spherical harmonic transforms and frequency domain convolutions of spherical images. This tool can also be used to detect multi-scale features on geometric surfaces. When we have a point cloud of a surface of a particular class of object, whether generated by stereo reconstruction or by some other modality, we can use statistics and machine learning to more robustly estimate the surface. If we have at our disposal a database of surfaces of a particular type of object, such as the human face, we can compute statistics over this database to constrain the possible shape a new surface of this type can take. We show how a statistical spherical wavelet shape prior can be used to efficiently and robustly reconstruct a face shape from noisy point cloud data, including stereo data.
Subjects/Keywords: multi-scale;
wavelets;
stereo reconstruction;
omnidirectional vision;
real-time novel view synthesis;
real-time virtual walkthroughs;
spherical parameterizations;
spherical harmonics;
GPU programming
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Brunton, A. P. (2012). Multi-scale Methods for Omnidirectional Stereo with Application to Real-time Virtual Walkthroughs
. (Thesis). University of Ottawa. Retrieved from http://hdl.handle.net/10393/23552
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Brunton, Alan P. “Multi-scale Methods for Omnidirectional Stereo with Application to Real-time Virtual Walkthroughs
.” 2012. Thesis, University of Ottawa. Accessed December 14, 2019.
http://hdl.handle.net/10393/23552.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Brunton, Alan P. “Multi-scale Methods for Omnidirectional Stereo with Application to Real-time Virtual Walkthroughs
.” 2012. Web. 14 Dec 2019.
Vancouver:
Brunton AP. Multi-scale Methods for Omnidirectional Stereo with Application to Real-time Virtual Walkthroughs
. [Internet] [Thesis]. University of Ottawa; 2012. [cited 2019 Dec 14].
Available from: http://hdl.handle.net/10393/23552.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Brunton AP. Multi-scale Methods for Omnidirectional Stereo with Application to Real-time Virtual Walkthroughs
. [Thesis]. University of Ottawa; 2012. Available from: http://hdl.handle.net/10393/23552
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

University of Oxford
19.
Woodford, Oliver J.
Priors for new view synthesis.
Degree: PhD, 2009, University of Oxford
URL: http://ora.ox.ac.uk/objects/uuid:38962bda-6f0f-4158-89cf-8c641ebac486
;
https://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.580856
► New view synthesis (NVS) is the problem of generating a novel image of a scene, given a set of calibrated input images of the scene,…
(more)
▼ New view synthesis (NVS) is the problem of generating a novel image of a scene, given a set of calibrated input images of the scene, i.e. their viewpoints, and also that of the output image, are known. The problem is generally ill-posed – a large number of scenes can generate a given set of images, therefore there may be many equally likely (given the input data) output views. Some of these views will look less natural to a human observer than others, so prior knowledge of natural scenes is required to ensure that the result is visually plausible. The aim of this thesis is to compare and improve upon the various Markov random field} and conditional random field prior models, and their associated maximum a posteriori optimization frameworks, that are currently the state of the art for NVS and stereo (itself a means to NVS). A hierarchical example-based image prior is introduced which, when combined with a multi-resolution framework, accelerates inference by an order of magnitude, whilst also improving the quality of rendering. A parametric image prior is tested using a number of novel discrete optimization algorithms. This general prior is found to be less well suited to the NVS problem than sequence-specific priors, generating two forms of undesirable artifact, which are discussed. A novel pairwise clique image prior is developed, allowing inference using powerful optimizers. The prior is shown to perform better than a range of other pairwise image priors, distinguishing as it does between natural and artificial texture discontinuities. A dense stereo algorithm with geometrical occlusion model is converted to the task of NVS. In doing so, a number of challenges are novelly addressed; in particular, the new pairwise image prior is employed to align depth discontinuities with genuine texture edges in the output image. The resulting joint prior over smoothness and texture is shown to produce cutting edge rendering performance. Finally, a powerful new inference framework for stereo that allows the tractable optimization of second order smoothness priors is introduced. The second order priors are shown to improve reconstruction over first order priors in a number of situations.
Subjects/Keywords: 519.2; Information engineering; Image understanding; Robotics; prior model; new view synthesis; optimization; markov random field
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Woodford, O. J. (2009). Priors for new view synthesis. (Doctoral Dissertation). University of Oxford. Retrieved from http://ora.ox.ac.uk/objects/uuid:38962bda-6f0f-4158-89cf-8c641ebac486 ; https://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.580856
Chicago Manual of Style (16th Edition):
Woodford, Oliver J. “Priors for new view synthesis.” 2009. Doctoral Dissertation, University of Oxford. Accessed December 14, 2019.
http://ora.ox.ac.uk/objects/uuid:38962bda-6f0f-4158-89cf-8c641ebac486 ; https://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.580856.
MLA Handbook (7th Edition):
Woodford, Oliver J. “Priors for new view synthesis.” 2009. Web. 14 Dec 2019.
Vancouver:
Woodford OJ. Priors for new view synthesis. [Internet] [Doctoral dissertation]. University of Oxford; 2009. [cited 2019 Dec 14].
Available from: http://ora.ox.ac.uk/objects/uuid:38962bda-6f0f-4158-89cf-8c641ebac486 ; https://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.580856.
Council of Science Editors:
Woodford OJ. Priors for new view synthesis. [Doctoral Dissertation]. University of Oxford; 2009. Available from: http://ora.ox.ac.uk/objects/uuid:38962bda-6f0f-4158-89cf-8c641ebac486 ; https://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.580856
20.
Mora, Elie-Gabriel.
Codage multi-vues multi-profondeur pour de nouveaux services multimédia : Multiview video plus depth coding for new multimedia services.
Degree: Docteur es, Signal et images, 2014, Paris, ENST
URL: http://www.theses.fr/2014ENST0007
► Les travaux effectués durant cette thèse de doctorat ont pour but d’augmenter l’efficacité de codage dans 3D-HEVC. Nous proposons des approches conventionnelles orientées vers la…
(more)
▼ Les travaux effectués durant cette thèse de doctorat ont pour but d’augmenter l’efficacité de codage dans 3D-HEVC. Nous proposons des approches conventionnelles orientées vers la normalisation vidéo, ainsi que des approches en rupture basées sur le flot optique. En approches conventionnelles, nous proposons une méthode qui prédit les modes Intra de profondeur avec ceux de texture. L’héritage est conditionné par un critère qui mesure le degré de similitude entre les deux modes. Ensuite, nous proposons deux méthodes pour améliorer la prédiction inter-vue du mouvement dans 3D-HEVC. La première ajoute un vecteur de disparité comme candidat inter-vue dans la liste des candidats du Merge, et la seconde modifie le processus de dérivation de ce vecteur. Finalement, un outil de codage intercomposantes est proposé, où le lien entre les arbres quaternaires de texture et de profondeur est exploité pour réduire le temps d’encodage et le débit, à travers un codage conjoint des deux arbres. Dans la catégorie des approches en rupture, nous proposons deux méthodes basées sur l’estimation de champs denses de vecteurs de mouvement en utilisant le flot optique. La première calcule un champ au niveau d’une vue de base reconstruite, puis l’extrapole au niveau d’une vue dépendante, où il est hérité par les unités de prédiction en tant que candidat dense du Merge. La deuxième méthode améliore la synthèse de vues : quatre champs sont calculés au niveau de deux vues de référence en utilisant deux références temporelles. Ils sont ensuite extrapolés au niveau d’une vue synthétisée et corrigés en utilisant une contrainte épipolaire. Les quatre prédictions correspondantes sont ensuite combinées.
This PhD. thesis deals with improving the coding efficiency in 3D-HEVC. We propose both constrained approaches aimed towards standardization, and also more innovative approaches based on optical flow. In the constrained approaches category, we first propose a method that predicts the depth Intra modes using the ones of the texture. The inheritance is driven by a criterion measuring how much the two are expected to match. Second, we propose two simple ways to improve inter-view motion prediction in 3D-HEVC. The first adds an inter-view disparity vector candidate in the Merge list and the second modifies the derivation process of this disparity vector. Third, an inter-component tool is proposed where the link between the texture and depth quadtree structures is exploited to save both runtime and bits through a joint coding of the quadtrees. In the more innovative approaches category, we propose two methods that are based on a dense motion vector field estimation using optical flow. The first computes such a field on a reconstructed base view. It is then warped at the level of a dependent view where it is inserted as a dense candidate in the Merge list of prediction units in that view. The second method improves the view synthesis process: four fields are computed at the level of the left and right reference views using a past and a future temporal…
Advisors/Committee Members: Pesquet-Popescu, Béatrice (thesis director), Cagnazzo, Marco (thesis director).
Subjects/Keywords: 3D-HEVC; Synthèse de vues; Flot optique; Liste des candidats du Merge; Initialisation et limitation d’un arbre quaternaire; Vecteur de disparité; Vecteur de mouvement; Mode Intra; 3D-HEVC; View synthesis; Optical flow; Merge candidate list; Quadtree initialization and limitation; Disparity vector; Motion vector; Intra mode
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Mora, E. (2014). Codage multi-vues multi-profondeur pour de nouveaux services multimédia : Multiview video plus depth coding for new multimedia services. (Doctoral Dissertation). Paris, ENST. Retrieved from http://www.theses.fr/2014ENST0007
Chicago Manual of Style (16th Edition):
Mora, Elie-Gabriel. “Codage multi-vues multi-profondeur pour de nouveaux services multimédia : Multiview video plus depth coding for new multimedia services.” 2014. Doctoral Dissertation, Paris, ENST. Accessed December 14, 2019.
http://www.theses.fr/2014ENST0007.
MLA Handbook (7th Edition):
Mora, Elie-Gabriel. “Codage multi-vues multi-profondeur pour de nouveaux services multimédia : Multiview video plus depth coding for new multimedia services.” 2014. Web. 14 Dec 2019.
Vancouver:
Mora E. Codage multi-vues multi-profondeur pour de nouveaux services multimédia : Multiview video plus depth coding for new multimedia services. [Internet] [Doctoral dissertation]. Paris, ENST; 2014. [cited 2019 Dec 14].
Available from: http://www.theses.fr/2014ENST0007.
Council of Science Editors:
Mora E. Codage multi-vues multi-profondeur pour de nouveaux services multimédia : Multiview video plus depth coding for new multimedia services. [Doctoral Dissertation]. Paris, ENST; 2014. Available from: http://www.theses.fr/2014ENST0007

Wright State University
21.
Robbeloth, Michael Christopher.
Recognition of Incomplete Objects based on Synthesis of
Views Using a Geometric Based Local-Global Graphs.
Degree: PhD, Computer Science and Engineering PhD, 2019, Wright State University
URL: http://rave.ohiolink.edu/etdc/view?acc_num=wright1557509373174391
► The recognition of single objects is an old research field with many techniques and robust results. The probabilistic recognition of incomplete objects, however, remains an…
(more)
▼ The recognition of single objects is an old research
field with many techniques and robust results. The probabilistic
recognition of incomplete objects, however, remains an active field
with challenging issues associated to shadows, illumination and
other visual characteristics. With object incompleteness, we mean
missing parts of a known object and not low-resolution images of
that object. The employment of various single machine-learning
methodologies for accurate classification of the incomplete objects
did not provide a robust answer to the challenging problem. In this
dissertation, we present a suite of high-level, model-based
computer vision techniques encompassing both geometric and machine
learning approaches to generate probabilistic matches of objects
with varying degrees and forms of non-deformed incompleteness. The
recognition of incomplete objects requires the formulation of a
database of six sided views (e.g., model) of an object from which
an identification can be made. The images are preprocessed (K-means
segmentation, and region growing code to generate fully defined
region and segment image information) from which local and global
geometric and characteristic properties are generated in a process
known as the Local-Global (L-G) Graph method. The characteristic
properties are then stored into a database for processing against
sample images featuring various types of missing features. The
sample images are then characterized in the same manner. After
this, a suite of methodologies is employed to match a sample
against an exemplar image in a multithreaded manner. The
approaches, which work with the multi-
view model database
characteristics in a parallel (e.g, multithreaded manner) determine
probabilistically by application of weighted outcomes the
application of various matching routines. These routines include
treating segment border regions as chain codes which are then
processed using various string matching algorithms, the matching by
center of moments from global graph construction, the matching of
chain code starting segment location, the differences in angles in
the center of moments between the model and sample images to find
the most similar graphs (e.g., image), and the use of Delaunay
triangulations of the center of moments formed during global graph
construction. The ability to find a most probable match is
extensible in the future to adding additional detection methods
with the appropriate weight adjustments. To enhance the detection
of incomplete objects, separate investigations have been made into
rotating the exemplars in standard increments and by object
extraction of segment border regions’ chain codes and subsequent
synthesis of objects from the multi-
view database. This approach is
novel and potentially extensible to compositing across multi-
view
segmented regions at the borders between views by either human
aided input of border relations or a systematic, automated
evaluation of common border objects between the views of an
exemplar. The first results are promising and trigger…
Advisors/Committee Members: Bourbakis, Nikolaos G. (Advisor).
Subjects/Keywords: Computer Science; incomplete objects; obstruction; Local-Global L-G Graph; synthesis of views; recognition; six-sided views; multi-view; chain code; geometric; segmentation; computer vision; image processing; graph algorithm; machine learning; Delaunay triangulations
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Robbeloth, M. C. (2019). Recognition of Incomplete Objects based on Synthesis of
Views Using a Geometric Based Local-Global Graphs. (Doctoral Dissertation). Wright State University. Retrieved from http://rave.ohiolink.edu/etdc/view?acc_num=wright1557509373174391
Chicago Manual of Style (16th Edition):
Robbeloth, Michael Christopher. “Recognition of Incomplete Objects based on Synthesis of
Views Using a Geometric Based Local-Global Graphs.” 2019. Doctoral Dissertation, Wright State University. Accessed December 14, 2019.
http://rave.ohiolink.edu/etdc/view?acc_num=wright1557509373174391.
MLA Handbook (7th Edition):
Robbeloth, Michael Christopher. “Recognition of Incomplete Objects based on Synthesis of
Views Using a Geometric Based Local-Global Graphs.” 2019. Web. 14 Dec 2019.
Vancouver:
Robbeloth MC. Recognition of Incomplete Objects based on Synthesis of
Views Using a Geometric Based Local-Global Graphs. [Internet] [Doctoral dissertation]. Wright State University; 2019. [cited 2019 Dec 14].
Available from: http://rave.ohiolink.edu/etdc/view?acc_num=wright1557509373174391.
Council of Science Editors:
Robbeloth MC. Recognition of Incomplete Objects based on Synthesis of
Views Using a Geometric Based Local-Global Graphs. [Doctoral Dissertation]. Wright State University; 2019. Available from: http://rave.ohiolink.edu/etdc/view?acc_num=wright1557509373174391
.