Interactive Focus+Context Rendering for Hexahedral Mesh Inspection

Christoph Neuhauser, Junpeng Wang, Rüdiger Westermann

View presentation: 2021-10-28T15:00:00Z GMT-0600 Change your timezone on the schedule page
2021-10-28T15:00:00Z
Exemplar figure, described by caption below
In the image, the visualization of a hexahedral mesh using our Focus+Context rendering technique can be seen. Coarse mesh edges are blended into a volumetric representation of the mesh. Cell deformations are mapped to color and opacity. Users can interactively explore the hex-mesh by moving a lens-like hotspot. In this lens region, more detailed edge structures or the boundary surface of highly deformed cells can be shown.
Fast forward

Direct link to video on YouTube: https://youtu.be/SjM4kUzl8Xw

Keywords

Visualization of Hex-Meshes, Real-Time Rendering, GPUs

Abstract

The visual inspection of a hexahedral mesh with respect to element quality is difficult due to clutter and occlusions that are produced when rendering all element faces or their edges simultaneously. Current approaches overcome this problem by using focus on specific elements that are then rendered opaque, and carving away all elements occluding their view. In this work, we make use of advanced GPU shader functionality to generate a focus+context rendering that highlights the elements in a selected region and simultaneously conveys the global mesh structure and deformation field. To achieve this, we propose a gradual transition from edge-based focus rendering to volumetric context rendering, by combining fragment shader-based edge and face rendering with per-pixel fragment lists. A fragment shader smoothly transitions between wireframe and face-based rendering, including focus-dependent rendering style and depth-dependent edge thickness and halos, and per-pixel fragment lists are used to blend fragments in correct visibility order. To maintain the global mesh structure in the context regions, we propose a new method to construct a sheet-based level-of-detail hierarchy and smoothly blend it with volumetric information. The user guides the exploration process by moving a lens-like hotspot. Since all operations are performed on the GPU, interactive frame rates are achieved even for large meshes.