Walton, DR;
Thomas, D;
Steed, AJ;
Sugimoto, A;
(2017)
Synthesis of environment maps for mixed reality.
In:
Mixed and Augmented Reality (ISMAR), 2017 IEEE International Symposium on.
IEEE
Preview |
Text
ISMAR_RevisedSubmission_2017_06.pdf - Accepted Version Download (12MB) | Preview |
Abstract
When rendering virtual objects in a mixed reality application, it is helpful to have access to an environment map that captures the appearance of the scene from the perspective of the virtual object. It is straightforward to render virtual objects into such maps, but capturing and correctly rendering the real components of the scene into the map is much more challenging. This information is often recovered from physical light probes, such as reflective spheres or fisheye cameras, placed at the location of the virtual object in the scene. For many application areas, however, real light probes would be intrusive or impractical. Ideally, all of the information necessary to produce detailed environment maps could be captured using a single device. We introduce a method using an RGBD camera and a small fisheye camera, contained in a single unit, to create environment maps at any location in an indoor scene. The method combines the output from both cameras to correct for their limited field of view and the displacement from the virtual object, producing complete environment maps suitable for rendering the virtual content in real time. Our method improves on previous probeless approaches by its ability to recover high-frequency environment maps. We demonstrate how this can be used to render virtual objects which shadow, reflect and refract their environment convincingly.
Type: | Proceedings paper |
---|---|
Title: | Synthesis of environment maps for mixed reality |
Event: | IEEE International Symposium on Mixed and Augmented Reality |
Location: | Nantes, France |
Dates: | 09 October 2017 - 13 October 2017 |
ISBN-13: | 978-1-5386-2944-4 |
Open access status: | An open access version is available from UCL Discovery |
DOI: | 10.1109/ISMAR.2017.24 |
Publisher version: | http://dx.doi.org/978-1-5386-2944-4 |
Language: | English |
Additional information: | This version is the author accepted manuscript. For information on re-use, please refer to the publisher’s terms and conditions. |
Keywords: | Cameras, Probes, Lighting, Rendering (computer graphics), Real-time systems, Light sources, Virtual reality |
UCL classification: | UCL UCL > Provost and Vice Provost Offices > UCL BEAMS UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science > Dept of Computer Science UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Maths and Physical Sciences UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Maths and Physical Sciences > Dept of Space and Climate Physics |
URI: | https://discovery-pp.ucl.ac.uk/id/eprint/1569655 |
Archive Staff Only
![]() |
View Item |