Downloads provided by UsageCounts
ULB ToysTable sequence by LISA ULB The test sequence "ULB ToysTable" is provided by Daniele Bonatto, Sarah Fachada, Gauthier Lafruit, members of the LISA department, EPB (Ecole Polytechnique de Bruxelles), ULB (Universite Libre de Bruxelles), Belgium. License: CC BY-NC-SA Terms of Use: Anykind of publication or report using this sequence should refer to the following references. [1] Daniele Bonatto, Sarah Fachada, Gauthier Lafruit, "ULB ToysTable", 2021. @misc{bonatto_toystable_2021, title = {{ULB} {ToysTable}}, author = {Bonatto, Daniele and Fachada, Sarah and Lafruit, Gauthier}, month = feb, year = {2021}, doi = {10.5281/zenodo.5055542} } [2] A. Schenkel, D. Bonatto, S. Fachada, H.-L. Guillaume, et G. Lafruit, « Natural Scenes Datasets for Exploration in 6DOF Navigation », in 2018 International Conference on 3D Immersion (IC3D), Brussels, Belgium, déc. 2018, p. 1-8. doi: 10.1109/IC3D.2018.8657865. @inproceedings{schenkel_natural_b_2018, address = {Brussels, Belgium}, title = {Natural {Scenes} {Datasets} for {Exploration} in {6DOF} {Navigation}}, isbn = {978-1-5386-7590-8}, url = {https://doi.org/10.1109/IC3D.2018.8657865}, doi = {10.1109/IC3D.2018.8657865}, language = {en}, urldate = {2019-04-11}, booktitle = {2018 {International} {Conference} on {3D} {Immersion} ({IC3D})}, publisher = {IEEE}, author = {Schenkel, Arnaud and Bonatto, Daniele and Fachada, Sarah and Guillaume, Henry-Louis and Lafruit, Gauthier}, month = dec, year = {2018}, pages = {1--8} } [3] D. Bonatto, A. Schenkel, T. Lenertz, Y. Li, et G. Lafruit, « [MPEG-I Visual] ULB High Density 2D/3D Camera Array data set, version 2 [m41083] », in ISO/IEC JTC1/SC29/WG11 MPEG2017/M41083, Torino, Italy, juill. 2017. @inproceedings{bonatto_mpeg-i_2017, address = {Torino, Italy}, title = {[{MPEG}-{I} {Visual}] {ULB} {High} {Density} {2D}/{3D} {Camera} {Array} data set, version 2 [m41083]}, doi = {ISO/IEC JTC1/SC29/WG11 MPEG2017/M41083}, author = {Bonatto, Daniele and Schenkel, Arnaud and Lenertz, Tim and Li, Yan and Lafruit, Gauthier}, month = jul, year = {2017} } Production: Laboratory of Image Synthesis and Analysis, LISA department, Ecole Polytechnique de Bruxelles, Universite Libre de Bruxelles, Belgium. Content: This dataset contains a static test scene created using a robotic bench described in [3]. We provide RGB textures and their associated depth maps captured using a Microsoft Kinect v2. We also provide depth maps estimated using MPEG's Depth Estimation Reference Software (DERS) [5]. The scene contains a table with several toys, boxes, a chessboard and the datacolor Spydercheckr® 24. The pictures were taken in a controlled light environment. In a post-processing pass, the colors were corrected and the depth map undistorded and reprojected as described in [2] to match the RGB images. The dataset contains two bands of regurarly spaced 5x5 images (Plane A) and 3x5 images (Plane B) respectively. In addition to the images and their depth maps, an accurate camera calibration file is provided following the format of [4]. It was computed as described in [2]. The dataset contains: - a `camera.json` file in OMAF coordinates system (Camera position: X: forwards, Y:left, Z: up, Rotation: yaw, pitch, roll) [3], - a `textures` folder containing the rendered views in png format, - a `depths_DERS` folder containing the associated depth maps in exr format. References and links: [4] S. Fachada, B. Kroon, D. Bonatto, B. Sonneveldt, et G. Lafruit, « Reference View Synthesizer (RVS) 2.0 manual, [N17759] », juill. 2018. [5] S. Rogge and D. Bonatto and J. Sancho and R. Salvador and E. Juarez and A. Munteanu and G. Lafruit, "MPEG-I Depth Estimation Reference Software", in 2019 International Conference on 3D Immersion (IC3D), 2019.
Camera Array, Depth Image-Based Rendering, View Synthesis
Camera Array, Depth Image-Based Rendering, View Synthesis
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
| views | 102 | |
| downloads | 65 |

Views provided by UsageCounts
Downloads provided by UsageCounts