@article{MATTAUSCH-2015-CHCRT, title = "CHC+RT: Coherent Hierarchical Culling for Ray Tracing", author = "Oliver Mattausch and Jir\'{i} Bittner and Alberto Jaspe and Enrico Gobbetti and Michael Wimmer and Renato Pajarola", year = "2015", abstract = "We propose a new technique for in-core and out-of-core GPU ray tracing using a generalization of hierarchical occlusion culling in the style of the CHC++ method. Our method exploits the rasterization pipeline and hardware occlusion queries in order to create coherent batches of work for localized shader-based ray tracing kernels. By combining hierarchies in both ray space and object space, the method is able to share intermediate traversal results among multiple rays. We exploit temporal coherence among similar ray sets between frames and also within the given frame. A suitable management of the current visibility state makes it possible to benefit from occlusion culling for less coherent ray types like diffuse reflections. Since large scenes are still a challenge for modern GPU ray tracers, our method is most useful for scenes with medium to high complexity, especially since our method inherently supports ray tracing highly complex scenes that do not fit in GPU memory. For in-core scenes our method is comparable to CUDA ray tracing and performs up to 5.94 × better than pure shader-based ray tracing.", month = may, journal = "Computer Graphics Forum", volume = "34", number = "2", issn = "1467-8659", pages = "537--548", keywords = "occlusion culling, ray tracing", URL = "https://www.cg.tuwien.ac.at/research/publications/2015/MATTAUSCH-2015-CHCRT/", } @article{preiner2014clop, title = "Continuous Projection for Fast L1 Reconstruction", author = "Reinhold Preiner and Oliver Mattausch and Murat Arikan and Renato Pajarola and Michael Wimmer", year = "2014", abstract = "With better and faster acquisition devices comes a demand for fast robust reconstruction algorithms, but no L1-based technique has been fast enough for online use so far. In this paper, we present a novel continuous formulation of the weighted locally optimal projection (WLOP) operator based on a Gaussian mixture describing the input point density. Our method is up to 7 times faster than an optimized GPU implementation of WLOP, and achieves interactive frame rates for moderately sized point clouds. We give a comprehensive quality analysis showing that our continuous operator achieves a generally higher reconstruction quality than its discrete counterpart. Additionally, we show how to apply our continuous formulation to spherical mixtures of normal directions, to also achieve a fast robust normal reconstruction. Project Page: https://www.cg.tuwien.ac.at/~preiner/projects/clop/", month = aug, journal = "ACM Transactions on Graphics (Proc. of ACM SIGGRAPH 2014)", volume = "33", number = "4", issn = "0730-0301", doi = "10.1145/2601097.2601172", pages = "47:1--47:13", keywords = "point set, Gaussian mixture, Hierarchical EM, upsampling, dynamic reconstruction, L1 reconstruction", URL = "https://www.cg.tuwien.ac.at/research/publications/2014/preiner2014clop/", } @article{hecher-2014-MH, title = "A Comparative Perceptual Study of Soft Shadow Algorithms", author = "Michael Hecher and Matthias Bernhard and Oliver Mattausch and Daniel Scherzer and Michael Wimmer", year = "2014", abstract = "We performed a perceptual user study of algorithms that approximate soft shadows in real time. Although a huge body of soft-shadow algorithms have been proposed, to our knowledge this is the first methodical study for comparing different real-time shadow algorithms with respect to their plausibility and visual appearance. We evaluated soft-shadow properties like penumbra overlap with respect to their relevance to shadow perception in a systematic way, and we believe that our results can be useful to guide future shadow approaches in their methods of evaluation. In this study, we also capture the predominant case of an inexperienced user observing shadows without comparing to a reference solution, such as when watching a movie or playing a game. One important result of this experiment is to scientifically verify that real-time soft-shadow algorithms, despite having become physically based and very realistic, can nevertheless be intuitively distinguished from a correct solution by untrained users.", month = jun, issn = "1544-3558", journal = "ACM Transactions on Applied Perception", number = "5", volume = "11", pages = "5:1--5:21", keywords = "Perception Studies, Soft Shadows", URL = "https://www.cg.tuwien.ac.at/research/publications/2014/hecher-2014-MH/", } @article{MATTAUSCH-2013-FSBE, title = "Freeform Shadow Boundary Editing", author = "Oliver Mattausch and Takeo Igarashi and Michael Wimmer", year = "2013", abstract = "We present an algorithm for artistically modifying physically based shadows. With our tool, an artist can directly edit the shadow boundaries in the scene in an intuitive fashion similar to freeform curve editing. Our algorithm then makes these shadow edits consistent with respect to varying light directions and scene configurations, by creating a shadow mesh from the new silhouettes. The shadow mesh helps a modified shadow volume algorithm cast shadows that conform to the artistic shadow boundary edits, while providing plausible interaction with dynamic environments, including animation of both characters and light sources. Our algorithm provides significantly more fine-grained local and direct control than previous artistic light editing methods, which makes it simple to adjust the shadows in a scene to reach a particular effect, or to create interesting shadow shapes and shadow animations. All cases are handled with a single intuitive interface, be it soft shadows, or (self-)shadows on arbitrary receivers.", month = may, journal = "Computer Graphics Forum (Proceeding of EUROGRAPHICS 2013)", volume = "32", number = "2", issn = "0167-7055", pages = "175--184", keywords = "shadows, real-time rendering", URL = "https://www.cg.tuwien.ac.at/research/publications/2013/MATTAUSCH-2013-FSBE/", } @article{SCHERZER-2012-TCM, title = "Temporal Coherence Methods in Real-Time Rendering", author = "Daniel Scherzer and Lei Yang and Oliver Mattausch and Diego Nehab and Pedro V. Sander and Michael Wimmer and Elmar Eisemann", year = "2012", abstract = "Nowadays, there is a strong trend towards rendering to higher-resolution displays and at high frame rates. This development aims at delivering more detail and better accuracy, but it also comes at a significant cost. Although graphics cards continue to evolve with an ever-increasing amount of computational power, the speed gain is easily counteracted by increasingly complex and sophisticated shading computations. For real-time applications, the direct consequence is that image resolution and temporal resolution are often the first candidates to bow to the performance constraints (e.g., although full HD is possible, PS3 and XBox often render at lower resolutions). In order to achieve high-quality rendering at a lower cost, one can exploit temporal coherence (TC). The underlying observation is that a higher resolution and frame rate do not necessarily imply a much higher workload, but a larger amount of redundancy and a higher potential for amortizing rendering over several frames. In this survey, we investigate methods that make use of this principle and provide practical and theoretical advice on how to exploit temporal coherence for performance optimization. These methods not only allow incorporating more computationally intensive shading effects into many existing applications, but also offer exciting opportunities for extending high-end graphics applications to lower-spec consumer-level hardware. To this end, we first introduce the notion and main concepts of TC, including an overview of historical methods. We then describe a general approach, image-space reprojection, with several implementation algorithms that facilitate reusing shading information across adjacent frames. We also discuss data-reuse quality and performance related to reprojection techniques. Finally, in the second half of this survey, we demonstrate various applications that exploit TC in real-time rendering. ", month = dec, issn = "1467-8659", journal = "Computer Graphics Forum", number = "8", volume = "31", pages = "2378--2408", keywords = "remote rendering; sampling, perception-based rendering, occlusion culling, non-photo-realistic rendering, level-of-detail, large data visualization, image-based rendering, global illumination, frame interpolation, anti-aliasing, shadows, streaming, temporal coherance, upsampling", URL = "https://www.cg.tuwien.ac.at/research/publications/2012/SCHERZER-2012-TCM/", } @article{knecht_martin_2012_RSMR, title = "Reciprocal Shading for Mixed Reality", author = "Martin Knecht and Christoph Traxler and Oliver Mattausch and Michael Wimmer", year = "2012", abstract = "In this paper we present a novel plausible rendering method for mixed reality systems, which is useful for many real-life application scenarios, like architecture, product visualization or edutainment. To allow virtual objects to seamlessly blend into the real environment, the real lighting conditions and the mutual illumination effects between real and virtual objects must be considered, while maintaining interactive frame rates. The most important such effects are indirect illumination and shadows cast between real and virtual objects. Our approach combines Instant Radiosity and Differential Rendering. In contrast to some previous solutions, we only need to render the scene once in order to find the mutual effects of virtual and real scenes. In addition, we avoid artifacts like double shadows or inconsistent color bleeding which appear in previous work. The dynamic real illumination is derived from the image stream of a fish-eye lens camera. The scene gets illuminated by virtual point lights, which use imperfect shadow maps to calculate visibility. A sufficiently fast scene reconstruction is done at run-time with Microsoft's Kinect sensor. Thus a time-consuming manual pre-modeling step of the real scene is not necessary. Our results show that the presented method highly improves the illusion in mixed-reality applications and significantly diminishes the artificial look of virtual objects superimposed onto real scenes.", month = nov, issn = "0097-8493", journal = "Computers & Graphics", number = "7", volume = "36", pages = "846--856", keywords = "Differential rendering, Reconstruction, Instant radiosity, Microsoft Kinect, Real-time global illumination, Mixed reality", URL = "https://www.cg.tuwien.ac.at/research/publications/2012/knecht_martin_2012_RSMR/", } @inproceedings{SCHWAERZLER-2012-FAS, title = "Fast Accurate Soft Shadows with Adaptive Light Source Sampling", author = "Michael Schw\"{a}rzler and Oliver Mattausch and Daniel Scherzer and Michael Wimmer", year = "2012", abstract = "Physically accurate soft shadows in 3D applications can be simulated by taking multiple samples from all over the area light source and accumulating them. Due to the unpredictability of the size of the penumbra regions, the required sampling density has to be high in order to guarantee smooth shadow transitions in all cases. Hence, several hundreds of shadow maps have to be evaluated in any scene configuration, making the process computationally expensive. Thus, we suggest an adaptive light source subdivision approach to select the sampling points adaptively. The main idea is to start with a few samples on the area light, evaluating there differences using hardware occlusion queries, and adding more sampling points if necessary. Our method is capable of selecting and rendering only the samples which contribute to an improved shadow quality, and hence generate shadows of comparable quality and accuracy. Even though additional calculation time is needed for the comparison step, this method saves valuable rendering time and achieves interactive to real-time frame rates in many cases where a brute force sampling method does not. ", month = nov, isbn = "978-3-905673-95-1", publisher = "Eurographics Association", location = "Magdeburg, Germany", booktitle = "Proceedings of the 17th International Workshop on Vision, Modeling, and Visualization (VMV 2012)", pages = "39--46", keywords = "soft shadows, real-time rendering", URL = "https://www.cg.tuwien.ac.at/research/publications/2012/SCHWAERZLER-2012-FAS/", } @article{MATTAUSCH-2012-TIS, title = "Tessellation-Independent Smooth Shadow Boundaries", author = "Oliver Mattausch and Daniel Scherzer and Michael Wimmer and Takeo Igarashi", year = "2012", abstract = "We propose an efficient and light-weight solution for rendering smooth shadow boundaries that do not reveal the tessellation of the shadow-casting geometry. Our algorithm reconstructs the smooth contours of the underlying mesh and then extrudes shadow volumes from the smooth silhouettes to render the shadows. For this purpose we propose an improved silhouette reconstruction using the vertex normals of the underlying smooth mesh. Then our method subdivides the silhouette loops until the contours are sufficiently smooth and project to smooth shadow boundaries. This approach decouples the shadow smoothness from the tessellation of the geometry and can be used to maintain equally high shadow quality for multiple LOD levels. It causes only a minimal change to the fill rate, which is the well-known bottleneck of shadow volumes, and hence has only small overhead. ", month = jun, journal = "Computer Graphics Forum", volume = "4", number = "31", issn = "1467-8659", pages = "1465--1470", keywords = "real-time rendering, shadows", URL = "https://www.cg.tuwien.ac.at/research/publications/2012/MATTAUSCH-2012-TIS/", } @incollection{MATTAUSCH-2012-EOV, title = "Efficient Online Visibility for Shadow Maps", author = "Oliver Mattausch and Jir\'{i} Bittner and Ari Silvennoinen and Daniel Scherzer and Michael Wimmer", year = "2012", abstract = "Standard online occlusion culling is able to vastly improve the rasterization performance of walkthrough applications by identifying large parts of the scene as invisible from the camera and rendering only the visible geometry. However, it is of little use for the acceleration of shadow map generation (i.e., rasterizing the scene from the light view [Williams 78]), so that typically a high percentage of the geometry will be visible when rendering shadow maps. For example, in outdoor scenes typical viewpoints are near the ground and therefore have significant occlusion, while light viewpoints are higher up and see most of the geometry. Our algorithm remedies this situation by quickly detecting and culling the geometry that does not contribute to the shadow in the final image.", month = feb, booktitle = "GPU Pro 3: Advanced Rendering Techniques", editor = "Wolfgang Engel", isbn = "978-1439887820", publisher = "CRC Press", keywords = "shadow maps, visibility culling", URL = "https://www.cg.tuwien.ac.at/research/publications/2012/MATTAUSCH-2012-EOV/", } @incollection{matt2011, title = "Temporal Screen-Space Ambient Occlusion", author = "Oliver Mattausch and Daniel Scherzer and Michael Wimmer", year = "2011", month = feb, booktitle = "GPU Pro 2", editor = "Wolfgang Engel", isbn = "978-1568817187", publisher = "A.K. Peters", keywords = "ambient occlusion", URL = "https://www.cg.tuwien.ac.at/research/publications/2011/matt2011/", } @incollection{2011scherzerE, title = "Fast Soft Shadows with Temporal Coherence", author = "Daniel Scherzer and Michael Schw\"{a}rzler and Oliver Mattausch", year = "2011", month = feb, booktitle = "GPU Pro 2", editor = "Wolfgang Engel", isbn = "978-1568817187", publisher = "A.K. Peters", URL = "https://www.cg.tuwien.ac.at/research/publications/2011/2011scherzerE/", } @inproceedings{bittner-2011-scc, title = "Shadow Caster Culling for Efficient Shadow Mapping", author = "Jir\'{i} Bittner and Oliver Mattausch and Ari Silvennoinen and Michael Wimmer", year = "2011", abstract = "We propose a novel method for efficient construction of shadow maps by culling shadow casters which do not contribute to visible shadows. The method uses a mask of potential shadow receivers to cull shadow casters using a hierarchical occlusion culling algorithm. We propose several variants of the receiver mask implementations with different culling efficiency and computational costs. For scenes with statically focused shadow maps we designed an efficient strategy to incrementally update the shadow map, which comes close to the rendering performance for unshadowed scenes. We show that our method achieves 3x-10x speedup for rendering large city like scenes and 1.5x-2x speedup for rendering an actual game scene.", month = feb, isbn = "978-1-4503-0565-5", publisher = "ACM", organization = "ACM SIGGRAPH", location = "San Francisco", booktitle = "Proceedings of the ACM SIGGRAPH Symposium on Interactive 3D Graphics and Games 2011", pages = "81--88", keywords = "occlusion culling, shadow mapping", URL = "https://www.cg.tuwien.ac.at/research/publications/2011/bittner-2011-scc/", } @inproceedings{scherzer2011c, title = "A Survey on Temporal Coherence Methods in Real-Time Rendering", author = "Daniel Scherzer and Lei Yang and Oliver Mattausch and Diego Nehab and Pedro V. Sander and Michael Wimmer and Elmar Eisemann", year = "2011", abstract = "Nowadays, there is a strong trend towards rendering to higher-resolution displays and at high frame rates. This development aims at delivering more detail and better accuracy, but it also comes at a significant cost. Although graphics cards continue to evolve with an ever-increasing amount of computational power, the processing gain is counteracted to a high degree by increasingly complex and sophisticated pixel computations. For real-time applications, the direct consequence is that image resolution and temporal resolution are often the first candidates to bow to the performance constraints (e.g., although full HD is possible, PS3 and XBox often render at lower resolutions). In order to achieve high-quality rendering at a lower cost, one can exploit emph{temporal coherence} (TC). The underlying observation is that a higher resolution and frame rate do not necessarily imply a much higher workload, but a larger amount of redundancy and a higher potential for amortizing rendering over several frames. In this STAR, we will investigate methods that make use of this principle and provide practical and theoretical advice on how to exploit temporal coherence for performance optimization. These methods not only allow us to incorporate more computationally intensive shading effects into many existing applications, but also offer exciting opportunities for extending high-end graphics applications to lower-spec consumer-level hardware. To this end, we first introduce the notion and main concepts of TC, including an overview of historical methods. We then describe a key data structure, the so-called emph{reprojection cache}, with several supporting algorithms that facilitate reusing shading information from previous frames. Its usefulness is illustrated in the second part of the STAR, where we present various applications. We illustrate how expensive pixel shaders, multi-pass shading effects, stereo rendering, shader antialiasing, shadow casting, and global-illumination effects can profit from pixel reuse. Furthermore, we will see that optimizations for visibility culling and object-space global illumination can also be achieved by exploiting TC. This STAR enables the reader to gain an overview of many techniques in this cutting-edge field and provides many insights into algorithmic choices and implementation issues. It delivers working knowledge of how various existing techniques are optimized via data reuse. Another goal of this STAR is to inspire the reader and to raise awareness for temporal coherence as an elegant tool that could be a crucial component to satisfy the recent need for higher resolution and more detailed content. ", booktitle = "EUROGRAPHICS 2011 State of the Art Reports", location = "Llandudno UK", publisher = "Eurographics Association", pages = "101--126", URL = "https://www.cg.tuwien.ac.at/research/publications/2011/scherzer2011c/", } @article{mattausch-2010-tao, title = "High-Quality Screen-Space Ambient Occlusion using Temporal Coherence", author = "Oliver Mattausch and Daniel Scherzer and Michael Wimmer", year = "2010", abstract = "Ambient occlusion is a cheap but effective approximation of global illumination. Recently, screen-space ambient occlusion (SSAO) methods, which sample the frame buffer as a discretization of the scene geometry, have become very popular for real-time rendering. We present temporal SSAO (TSSAO), a new algorithm which exploits temporal coherence to produce high-quality ambient occlusion in real time. Compared to conventional SSAO, our method reduces both noise as well as blurring artifacts due to strong spatial filtering, faithfully representing fine-grained geometric structures. Our algorithm caches and reuses previously computed SSAO samples, and adaptively applies more samples and spatial filtering only in regions that do not yet have enough information available from previous frames. The method works well for both static and dynamic scenes.", month = dec, issn = "0167-7055", journal = "Computer Graphics Forum", number = "8", volume = "29", pages = "2492--2503", keywords = "temporal coherence, ambient occlusion, real-time rendering", URL = "https://www.cg.tuwien.ac.at/research/publications/2010/mattausch-2010-tao/", } @inproceedings{knecht_martin_2010_DIR, title = "Differential Instant Radiosity for Mixed Reality", author = "Martin Knecht and Christoph Traxler and Oliver Mattausch and Werner Purgathofer and Michael Wimmer", year = "2010", abstract = "In this paper we present a novel plausible realistic rendering method for mixed reality systems, which is useful for many real life application scenarios, like architecture, product visualization or edutainment. To allow virtual objects to seamlessly blend into the real environment, the real lighting conditions and the mutual illumination effects between real and virtual objects must be considered, while maintaining interactive frame rates (20-30fps). The most important such effects are indirect illumination and shadows cast between real and virtual objects. Our approach combines Instant Radiosity and Differential Rendering. In contrast to some previous solutions, we only need to render the scene once in order to find the mutual effects of virtual and real scenes. The dynamic real illumination is derived from the image stream of a fish-eye lens camera. We describe a new method to assign virtual point lights to multiple primary light sources, which can be real or virtual. We use imperfect shadow maps for calculating illumination from virtual point lights and have significantly improved their accuracy by taking the surface normal of a shadow caster into account. Temporal coherence is exploited to reduce flickering artifacts. Our results show that the presented method highly improves the illusion in mixed reality applications and significantly diminishes the artificial look of virtual objects superimposed onto real scenes.", month = oct, note = "Best Paper Award!", location = "Seoul", booktitle = "Proceedings of the 2010 IEEE International Symposium on Mixed and Augmented Reality (ISMAR 2010)", pages = "99--107", keywords = "Instant Radiosity, Differential Rendering, Real-time Global Illumination, Mixed Reality", URL = "https://www.cg.tuwien.ac.at/research/publications/2010/knecht_martin_2010_DIR/", } @phdthesis{Mattausch-2010-vcr, title = "Visibility Computations for Real-Time Rendering in General 3D Environments", author = "Oliver Mattausch", year = "2010", abstract = "Visibility computations are essential operations in computer graphics, which are required for rendering acceleration in the form of visibility culling, as well as for computing realistic lighting. Visibility culling, which is the main focus of this thesis, aims to provide output sensitivity by sending only visible primitives to the hardware. Regardless of the rapid development of graphics hardware, it is of crucial importance for many applications like game development or architectural design, as the demands on the hardware regarding scene complexity increase accordingly. Solving the visibility problem has been an important research topic for many years, and countless methods have been proposed. Interestingly, there are still open research problems up to this day, and many algorithms are either impractical or only usable for specific scene configurations, preventing their widespread use. Visibility culling algorithms can be separated into algorithms for visibility preprocessing and online occlusion culling. Visibility computations are also required to solve complex lighting interactions in the scene, ranging from soft and hard shadows to ambient occlusion and full fledged global illumination. It is a big challenge to answer hundreds or thousands of visibility queries within a fraction of a second in order to reach real-time frame rates, which is one goal that we want to achieve in this thesis. The contribution of this thesis are four novel algorithms that provide solutions for efficient visibility interactions in order to achieve high-quality output-sensitive real-time rendering, and are general in the sense that they work with any kind of 3D scene configuration. First we present two methods dealing with the issue of automatically partitioning view space and object space into useful entities that are optimal for the subsequent visibility computations. Amazingly, this problem area was mostly ignored despite its importance, and view cells are mostly tweaked by hand in practice in order to reach optimal performance – a very time consuming task. The first algorithm specifically deals with the creation of an optimal view space partition into view cells using a cost heuristics and sparse visibility sampling. The second algorithm extends this approach to optimize both view space subdivision and object space subdivision simultaneously. Next we present a hierarchical online culling algorithm that eliminates most limitations of previous approaches, and is rendering engine friendly in the sense that it allows easy integration and efficient material sorting. It reduces the main problem of previous algorithms – the overhead due to many costly state changes and redundant hardware occlusion queries – to a minimum, obtaining up to three times speedup over previous work. At last we present an ambient occlusion algorithm which works in screen space, and show that high-quality shading with effectively hundreds of samples per pixel is possible in real time for both static and dynamic scenes by utilizing temporal coherence to reuse samples from previous frames.", month = apr, address = "Favoritenstrasse 9-11/E193-02, A-1040 Vienna, Austria", school = "Institute of Computer Graphics and Algorithms, Vienna University of Technology ", keywords = "3D rendering, real-time rendering, ambient occlusion, visibility, occlusion culling", URL = "https://www.cg.tuwien.ac.at/research/publications/2010/Mattausch-2010-vcr/", } @inproceedings{scherzer2010d, title = "Exploiting Temporal Coherence in Real-Time Rendering", author = "Daniel Scherzer and Lei Yang and Oliver Mattausch", year = "2010", abstract = "Temporal coherence (TC), the correlation of contents between adjacent rendered frames, exists across a wide range of scenes and motion types in practical real-time rendering. By taking advantage of TC, we can save redundant computation and improve the performance of many rendering tasks significantly with only a marginal decrease in quality. This not only allows us to incorporate more computationally intensive shading effects to existing applications, but also offers exciting opportunities of extending high-end graphics applications to reach lower-spec consumer-level hardware. This course aims to introduce participants to the concepts of TC, and provide them the working practical and theoretical knowledge to exploit TC in a variety of shading tasks. It begins with an introduction of the general notion of TC in rendering, as well as an overview of the recent developments in this field. Then it focuses on a key data structure - the reverse reprojection cache, which is the foundation of many applications. The course proceeds with a number of extensions of the basic algorithm for assisting in multi-pass shading effects, shader antialiasing, casting shadows and global-illumination effects. Finally, several more general coherence topics beyond pixel reuse are introduced, including visibility culling optimization and object-space global-illumination approximations. For all the major techniques and applications covered, implementation and practical issues involved in development are addressed in detail. In general, we emphasize "know how" and the guidelines related to algorithm choices. After the course, participants are encouraged to find and utilize TC in their own applications and rapidly adapt existing algorithms to meet their requirements.", booktitle = "SIGGRAPH Asia 2010 Courses", location = "Seoul, S\"{u}dkorea", keywords = "shadows, temporal coherence, real-time, rendering", URL = "https://www.cg.tuwien.ac.at/research/publications/2010/scherzer2010d/", } @book{mattausch-2010-var, title = "Visibility Algorithms for Real-Time Rendering", author = "Oliver Mattausch", year = "2010", isbn = "978-3-8381-1887-1", pages = "212", publisher = "Suedwestdeutscher Verlag fuer Hochschulschriften", keywords = "real-time rendering, occlusion culling , ambient occlusion, visibility", URL = "https://www.cg.tuwien.ac.at/research/publications/2010/mattausch-2010-var/", } @inproceedings{SSMW09, title = "Real-Time Soft Shadows Using Temporal Coherence", author = "Daniel Scherzer and Michael Schw\"{a}rzler and Oliver Mattausch and Michael Wimmer", year = "2009", abstract = "A vast amount of soft shadow map algorithms have been presented in recent years. Most use a single sample hard shadow map together with some clever filtering technique to calculate perceptually or even physically plausible soft shadows. On the other hand there is the class of much slower algorithms that calculate physically correct soft shadows by taking and combining many samples of the light. In this paper we present a new soft shadow method that combines the benefits of these approaches. It samples the light source over multiple frames instead of a single frame, creating only a single shadow map each frame. Where temporal coherence is low we use spatial filtering to estimate additional samples to create correct and very fast soft shadows. ", month = dec, isbn = "978-3642103308", series = "Lecture Notes in Computer Science", publisher = "Springer", location = "Las Vegas, Nevada, USA", editor = "Bebis, G.; Boyle, R.; Parvin, B.; Koracin, D.; Kuno, Y.; Wang, J.; Pajarola, R.; Lindstrom, P.; Hinkenjann, A.; Encarnacao, M.; Silva, C.; Coming, D.", booktitle = "Advances in Visual Computing: 5th International Symposium on Visual Computing (ISVC 2009)", pages = "13--24", keywords = "real-time rendering, soft shadows", URL = "https://www.cg.tuwien.ac.at/research/publications/2009/SSMW09/", } @article{BITTNER-2009-AGVS, title = "Adaptive Global Visibility Sampling", author = "Jir\'{i} Bittner and Oliver Mattausch and Peter Wonka and Vlastimil Havran and Michael Wimmer", year = "2009", abstract = "In this paper we propose a global visibility algorithm which computes from-region visibility for all view cells simultaneously in a progressive manner. We cast rays to sample visibility interactions and use the information carried by a ray for all view cells it intersects. The main contribution of the paper is a set of adaptive sampling strategies based on ray mutations that exploit the spatial coherence of visibility. Our method achieves more than an order of magnitude speedup compared to per-view cell sampling. This provides a practical solution to visibility preprocessing and also enables a new type of interactive visibility analysis application, where it is possible to quickly inspect and modify a coarse global visibility solution that is constantly refined. ", month = aug, journal = "ACM Transactions on Graphics", volume = "28", number = "3", issn = "0730-0301", pages = "94:1--94:10", keywords = "occlusion culling, visibility sampling, visibility, PVS", URL = "https://www.cg.tuwien.ac.at/research/publications/2009/BITTNER-2009-AGVS/", } @incollection{BITTNER-2009-GEFOC, title = "Game-Engine-Friendly Occlusion Culling", author = "Jir\'{i} Bittner and Oliver Mattausch and Michael Wimmer", year = "2009", abstract = "This article presents a method which minimizes the overhead associated with occlusion queries. The method reduces the number of required state changes and should integrate easily with most game engines. The key ideas are batching of the queries and interfacing with the game engine using a dedicated render queue. We also present some additional optimizations which reduce the number of queries issued as well as the number of rendered primitives. The algorithm is based on the well-known Coherent Hierarchical Culling algorithm.", month = mar, booktitle = "SHADERX7: Advanced Rendering Techniques", chapter = "8.3", editor = "Wolfang Engel", isbn = "1-58450-598-2", publisher = "Charles River Media", volume = "7", keywords = "real-time rendering, occlusion culling", URL = "https://www.cg.tuwien.ac.at/research/publications/2009/BITTNER-2009-GEFOC/", } @article{mattausch-2008-CHC, title = "CHC++: Coherent Hierarchical Culling Revisited", author = "Oliver Mattausch and Jir\'{i} Bittner and Michael Wimmer", year = "2008", abstract = "We present a new algorithm for efficient occlusion culling using hardware occlusion queries. The algorithm significantly improves on previous techniques by making better use of temporal and spatial coherence of visibility. This is achieved by using adaptive visibility prediction and query batching. As a result of the new optimizations the number of issued occlusion queries and the number of rendering state changes are significantly reduced. We also propose a simple method for determining tighter bounding volumes for occlusion queries and a method which further reduces the pipeline stalls. The proposed method provides up to an order of magnitude speedup over the previous state of the art. The new technique is simple to implement, does not rely on hardware calibration and integrates well with modern game engines.", month = apr, journal = "Computer Graphics Forum (Proceedings Eurographics 2008)", volume = "27", number = "2", issn = "0167-7055", pages = "221--230", keywords = "temporal coherence, dynamic occlusion culling, occlusion queries", URL = "https://www.cg.tuwien.ac.at/research/publications/2008/mattausch-2008-CHC/", } @inproceedings{MATTAUSCH-2007-OSP, title = "Optimized Subdivisions for Preprocessed Visibility", author = "Oliver Mattausch and Jir\'{i} Bittner and Peter Wonka and Michael Wimmer", year = "2007", abstract = "This paper describes a new tool for preprocessed visibility. It puts together view space and object space partitioning in order to control the render cost and memory cost of the visibility description generated by a visibility solver. The presented method progressively refines view space and object space subdivisions while minimizing the associated render and memory costs. Contrary to previous techniques, both subdivisions are driven by actual visibility information. We show that treating view space and object space together provides a powerful method for controlling the efficiency of the resulting visibility data structures.", month = may, isbn = "978-1-56881-337-0", publisher = "Canadian Human-Computer Communications Society", location = "Montreal, Canada", editor = "Christopher G. Healey and Edward Lank", booktitle = "Proceedings of Graphics Interface 2007", pages = "335--342", keywords = "visibility preprocessing, potentially visible sets, view cells", URL = "https://www.cg.tuwien.ac.at/research/publications/2007/MATTAUSCH-2007-OSP/", } @talk{mattausch-2007-iav, title = "1. Adaptive Visibility-Driven View Cell Construction", author = "Oliver Mattausch", year = "2007", abstract = "We present a new method for the automatic partitioning of view space into a multi-level view cell hierarchy. We use a cost-based model in order to minimize the average rendering time. Unlike previous methods, our model takes into account the actual visibility in the scene, and the partition is not restricted to planes given by the scene geometry. We show that the resulting view cell hierarchy works for different types of scenes and gives lower average rendering time than previously used methods.", event = "Konversatorium", location = "Prague University of Technology", keywords = "visibiility, view cells", URL = "https://www.cg.tuwien.ac.at/research/publications/2007/mattausch-2007-iav/", } @inproceedings{MATTAUSCH-2006-AVC, title = "Adaptive Visibility-Driven View Cell Construction", author = "Oliver Mattausch and Jir\'{i} Bittner and Michael Wimmer", year = "2006", abstract = "We present a new method for the automatic partitioning of view space into a multi-level view cell hierarchy. We use a cost-based model in order to minimize the average rendering time. Unlike previous methods, our model takes into account the actual visibility in the scene, and the partition is not restricted to planes given by the scene geometry. We show that the resulting view cell hierarchy works for different types of scenes and gives lower average rendering time than previously used methods.", month = jun, isbn = "3-90567-335-5", publisher = "Eurographics Association", organization = "Eurographics", location = "Nicosia, Cyprus", editor = "Wolfgang Heidrich and Tomas Akenine-Moller", booktitle = "Rendering Techniques 2006 (Proceedings Eurographics Symposium on Rendering)", pages = "195--206", keywords = "view cells, real-time rendering, visibility", URL = "https://www.cg.tuwien.ac.at/research/publications/2006/MATTAUSCH-2006-AVC/", } @mastersthesis{matt-masterthesis, title = "Practical Reconstruction Schemes and Hardware-Accelerated Direct Volume Rendering on Body-Centered Cubic Grids", author = "Oliver Mattausch", year = "2004", abstract = "It is well known in the signal-processing community that the Body-Centered Cubic grid is the optimal sampling grid in 3D. In volume visualization, the Cartesian grid is by far the most popular type of grid because it is convenient to handle. But it requires 29.3% more samples than the Body-Centered Cubic grid. In order to convince people used to Cartesian grids for years of the advantages of Body-Centered Cubic grids, we must prove their usability in many different volume rendering algorithms. Further we have to show that we get a performance gain without or with only slight loss of image quality compared to Cartesian grids. Therefore we introduce several practical reconstruction schemes on Body-Centered Cubic grids, which are very general and can be used in a number of applications and tasks. Together with the development of powerful and flexible consumer graphics hardware, interactive hardware-accelerated volume rendering algorithms gain popularity. Rendering performance becomes a big issue, which can be a strong argument in favour of Body-Centered Cubic grids. We adapted some of the most popular volume rendering approaches exploiting hardwareacceleration to Body-Centered Cubic grids: both 2D and 3D texture-based volume rendering and the projected tetrahedra algorithm. At least partly we succeeded in achieving a performance gain on our new grid and further produced some impressive rendering results comparable to the Cartesian grid version.", month = may, note = "1", address = "Favoritenstrasse 9-11/E193-02, A-1040 Vienna, Austria", school = "Institute of Computer Graphics and Algorithms, Vienna University of Technology ", URL = "https://www.cg.tuwien.ac.at/research/publications/2004/matt-masterthesis/", } @inproceedings{Mattausch-2003-Str, title = "Strategies for Interactive Exploration of 3D Flow Using Evenly-Spaced Illuminated Streamlines", author = "Oliver Mattausch and Thomas Theu{\ss}l and Helwig Hauser and Eduard Gr\"{o}ller", year = "2003", month = apr, isbn = "158113861x", publisher = "SCCG", editor = "K. Joy", booktitle = "Proceedings of Spring Conference on Computer Graphics", pages = "213 --222", keywords = "flow visualization, streamlines, focus and context", URL = "https://www.cg.tuwien.ac.at/research/publications/2003/Mattausch-2003-Str/", } @techreport{Mattausch-2003-SIE, title = "Strategies for Interactive Exploration of 3D Flow Using Evenly-Spaced Illuminated Streamlines", author = "Oliver Mattausch and Thomas Theu{\ss}l and Helwig Hauser and Eduard Gr\"{o}ller", year = "2003", abstract = "This paper presents several strategies to interactively explore 3D flow. Based on a fast illuminated streamlines algorithm, standard graphics hardware is sufficient to gain interactive rendering rates. Our approach does not require the user to have any prior knowledge of flow features. After the streamlines are computed in a short preprocessing time, the user can interactively change appearance and density of the streamlines to further explore the flow. Most important flow features like velocity or pressure not only can be mapped to all available streamline appearance properties like streamline width, material, opacity, but also to streamline density. To improve spatial perception of the 3D flow we apply techniques based on animation, depth cueing, and halos along a streamline if it is crossed by another streamline in the foreground. Finally, we make intense use of focus+context methods like magic volumes, region of interest driven streamline placing, and spotlights to solve the occlusion problem.", month = apr, number = "TR-186-2-03-04", address = "Favoritenstrasse 9-11/E193-02, A-1040 Vienna, Austria", institution = "Institute of Computer Graphics and Algorithms, Vienna University of Technology ", note = "human contact: technical-report@cg.tuwien.ac.at", keywords = "3D flow visualization, illuminated streamlines, interactive exploration, focus+context visualization", URL = "https://www.cg.tuwien.ac.at/research/publications/2003/Mattausch-2003-SIE/", } @studentproject{Mattausch-2003-ISL, title = "Evenly-Spaced Illuminated Streamlines in 3D", author = "Oliver Mattausch", year = "2003", abstract = "http://www.cg.tuwien.ac.at/courses/projekte_old/vis/finished/OMattausch/Flow3DWeb/index.htm", URL = "https://www.cg.tuwien.ac.at/research/publications/2003/Mattausch-2003-ISL/", } @techreport{Theussl-2002-RSH, title = "Reconstruction Schemes for High Quality Raycasting of the Body-Centered Cubic Grid", author = "Thomas Theu{\ss}l and Oliver Mattausch and Torsten M\"{o}ller and Eduard Gr\"{o}ller", year = "2002", abstract = "The body-centered cubic (BCC) grid has received attention in the volume visualization community recently due to its ability to represent the same data with almost 30\% fewer samples as compared to the Cartesian cubic (CC) grid. In this paper we present several resampling strategies for raycasting BCC grids. These strategies range from 2D interpolation in planes to piece-wise linear (barycentric) interpolation in a tetrahedral decomposition of the grid to trilinear and sheared trilinear interpolation. We compare them to raycasting with comparable resampling techniques in the commonly used CC grid in terms of computational complexity and visual quality.", month = dec, number = "TR-186-2-02-11", address = "Favoritenstrasse 9-11/E193-02, A-1040 Vienna, Austria", institution = "Institute of Computer Graphics and Algorithms, Vienna University of Technology ", note = "human contact: technical-report@cg.tuwien.ac.at", keywords = "interpolation, body-centered cubic grid, raycasting", URL = "https://www.cg.tuwien.ac.at/research/publications/2002/Theussl-2002-RSH/", }