EGPGV25: Eurographics Symposium on Parallel Graphics and Visualization
Permanent URI for this collection
Browse
Browsing EGPGV25: Eurographics Symposium on Parallel Graphics and Visualization by Subject "Volumetric models"
Now showing 1 - 2 of 2
Results Per Page
Sort Options
Item From Cluster to Desktop: A Cache-Accelerated INR framework for Interactive Visualization of Tera-Scale Data(The Eurographics Association, 2025) Zavorotny, Daniel; Wu, Qi; Bauer, David; Ma, Kwan-Liu; Reina, Guido; Rizzi, Silvio; Gueunet, CharlesMachine learning has enabled the use of implicit neural representations (INRs) to efficiently compress and reconstruct massive scientific datasets. However, despite advances in fast INR rendering algorithms, INR-based rendering remains computationally expensive, as computing data values from an INR is significantly slower than reading them from GPU memory. This bottleneck currently restricts interactive INR visualization to professional workstations. To address this challenge, we introduce an INR rendering framework accelerated by a scalable, multi-resolution GPU cache capable of efficiently representing tera-scale datasets. By minimizing redundant data queries and prioritizing novel volume regions, our method reduces the number of INR computations per frame, achieving an average 5× speedup over the state-of-the-art INR rendering method while still maintaining high visualization quality. Coupled with existing hardware-accelerated INR compressors, our framework enables scientists to generate and compress massive datasets in situ on high-performance computing platforms and then interactively explore them on consumer-grade hardware post hoc.Item Multi-Density Woodcock Tracking: Efficient & High-Quality Rendering for Multi-Channel Volumes(The Eurographics Association, 2025) Sahistan, Alper; Zellmann, Stefan; Morrical, Nate; Pascucci, Valerio; Wald, Ingo; Reina, Guido; Rizzi, Silvio; Gueunet, CharlesVolume rendering techniques for scientific visualization have increasingly transitioned toward Monte Carlo (MC) methods in recent years due to their flexibility and robustness. However, their application in multi-channel visualization remains underexplored. Traditional compositing-based approaches often employ arbitrary color blending functions, which lack a physical basis and can obscure data interpretation. We introduce multi-density Woodcock tracking, a simple and flexible extension of Woodcock tracking for multi-channel volume rendering that leverages the strengths of Monte Carlo methods to generate high-fidelity visuals. Our method offers a physically grounded solution for inter-channel color blending and eliminates the need for arbitrary blending functions. We also propose a unified blending modality by generalizing Woodcock's distance tracking method, facilitating seamless integration of alternative blending functions from prior works. Through evaluation across diverse datasets, we demonstrate that our approach maintains real-time interactivity while achieving high-quality visuals by accumulating frames over time.