The paper presents an effective method for calculating global illumination of scenes with a voxel representation of geometry, based on the Path Tracing method. The method is aimed to obtain a fast and reliable estimate of the luminance of global illumination in scenes obtained as a result of three-dimensional real-world scanning. To further increase the calculation efficiency, the method can be transferred to the GPU. The article discusses the main works on the visualization of volumetric geometry. The octree was chosen as the main representation of the geometry in memory, which allows accelerating ray tracing in the scene and reducing the amount of memory required. Optimizations of the ray tracing algorithm in octree and voxel matrix representations are described and applied. A method for estimating the calculation accuracy by evaluating the mean square error over the entire image is presented. Testing and comparison of the method results on two representations of geometry in memory: octree and three-dimensional matrix, as well as a comparison of the method with the original Path Tracing method, which revealed a twofold acceleration of the calculation, were carried out. Optimizations are proposed to further improve the efficiency of the method.
This research focuses on the possibility of building an alternative mixed reality (MR) system, which will eliminate all its main causes of visual discomfort and form a model of the real world in the virtual space, as much as possible corresponding to it. The relationship between virtual reality (VR) systems, which are limited to models of their own virtual world, and MR systems, which add virtual objects to the real world, is examined. This paper presents an approach based on generating a point cloud of static objects using RGB-D sensors of the MR device, its further classification and segmentation, and then searching for a similar CAD object in the appropriate database. The found virtual analogue, after appropriate transformations, replaces the real object of the scene. To create more realism, RGB images of real objects can be superimposed as textures on the corresponding virtual scene objects. The paper proposes a multimodal approach, which consists in searching for objects with similar modalities in databases. A virtual scene created in a single space using this approach eliminates the possibility of forming unnatural lighting and observation conditions for all objects, including virtual copies of real objects and added virtual objects.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.