2009
Permanent URI for this collection
Browse
Browsing 2009 by Issue Date
Now showing 1 - 20 of 25
Results Per Page
Sort Options
Item Feature Centric Volume Visualization(Malik, 11.12.2009) Malik, Muhammad MuddassirThis thesis presents techniques and algorithms for the effective exploration of volumetric datasets. The Visualization techniques are designed to focus on user specified features of interest. The proposed techniques are grouped into four chapters namely feature peeling, computation and visualization of fabrication artifacts, locally adaptive marching cubes, and comparative visualization for parameter studies of dataset series. The presented methods enable the user to efficiently explore the volumetric dataset for features of interest.Feature peeling is a novel rendering algorithm that analyzes ray profiles along lines of sight. The profiles are subdivided according to encountered peaks and valleys at so called transition points. The sensitivity of these transition points is calibrated via two thresholds. The slope threshold is based on the magnitude of a peak following a valley, while the peeling threshold measures the depth of the transition point relative to the neighboring rays. This technique separates the dataset into a number of feature layers.Fabrication artifacts are of prime importance for quality control engineers for first part inspection of industrial components. Techniques are presented in this thesis to measure fabrication artifacts through direct comparison of a reference CAD model with the corresponding industrial 3D X-ray computed tomography volume. Information from the CAD model is used to locate corresponding points in the volume data. Then various comparison metrics are computed to measure differences (fabrication artifacts) between the CAD model and the volumetric dataset. The comparison metrics are classified as either geometry-driven comparison techniques or visual-driven comparison techniques.The locally adaptive marching cubes algorithm is a modification of the marching cubes algorithm where instead of a global iso-value, each grid point has its own iso-value. This defines an iso-value field, which modifies the case identification process in the algorithm. An iso-value field enables the algorithm to correct biases within the dataset like low frequency noise, contrast drifts, local density variations, and other artifacts introduced by the measurement process. It can also be used for blending between different iso-surfaces (e.g., skin, and bone in a medical dataset).Comparative visualization techniques are proposed to carry out parameter studies for the special application area of dimensional measurement using industrial 3D X-ray computed tomography. A dataset series is generated by scanning a specimen multiple times by varying parameters of the scanning device. A high resolution series is explored using a planar reformatting based visualization system. A multi-image view and an edge explorer are proposed for comparing and visualizing gray values and edges of several datasets simultaneously. For fast data retrieval and convenient usability the datasets are bricked and efficient data structures are used.Item Uses of uncalibrated images to enrich3D models information(Dellepiane, 2009) Dellepiane, MatteoLa diminuzione dei costi delle fotocamere digitali semi-professionali ha portato allapossibilit per tutti di acquisire immagini ad alta definizione in modo molto semplice.Tuttavia, l interpretazione di queste immagini, nell ambito di tecniche di analisi dellascena di ricostruzione 3D della stessa, risulta ancora molto difficile a causa dellaricchezza di informazione acquisita. Nel caso in cui si conosca per una rappresentazione,anche semplificata, della scena, possibile estrarre dati interessanti in manieraautomatica o semi-automatica.Questi dati possono essere utilizzati in diversi modi per arricchire la qualit dei dati3D in possesso. Nell ambito di questa tesi, sono quindi presentate alcune tecnicheper l uso di immagini non registrate per l arricchimento di modelli 3D.In particolare, due possibili campi di applicazione sono considerati: l acquisizione,proiezione e visualizzazione dell informazione di colore e la modifica della geometriadi partenza.Per quanto riguarda la gestione del colore, sono presentate alcune soluzione praticheed efficaci, che hanno portato a importanti risultati nell ambito di svariati progettinell ambito in particolare dei Beni Culturali.Considerando invece le tecniche di modifica della geometria, sono presentati dueapprocci che apportano cambiamenti nella topologia di modelli 3D gi esistenti. Inparticolare, nella prima tecnica le informazioni estratte dalle immagini sono usateper produrre modelli tridimesionali di teste umane, usati per simulazioni di soundscattering nell ambito di applicazioni di 3D sound rendering. Il secondo metodo permetteinvece di completare modelli 3D con buchi, utilizzando immagini dell oggettoreale su cui sia stato proiettato un pattern laser predefinito.Infine, sono presentate alcune interessanti indicazioni a proposito di possibili sviluppifuturi dei metodi proposti, per delineare la direzione di questo promettente argomentodi ricerca. - The decrease in costs of semi-professional digital cameras has led to the possibilityfor everyone to acquire a very detailed description of a scene in a very short time.Unfortunately, the interpretation of the images is usually quite hard, due to the amountof data and the lack of robust and generic image analysis methods. Nevertheless, if ageometric description of the depicted scene is available, it gets much easier to extractinformation from 2D data.This information can be used to enrich the quality of the 3D data in several ways.In this thesis, several uses of sets of unregistered images for the enrichment of 3Dmodels are shown.In particular, two possible fields of application are presented: the color acquisition,projection and visualization and the geometry modification.Regarding color management, several practical and cheap solutions to overcome themain issues in this field are presented. Moreover, some real applications, mainly relatedto Cultural Heritage, show that provided methods are robust and effective.In the context of geometry modification, two approaches are presented to modify alreadyexisting 3D models. In the first one, information extracted from images is usedto deform a dummy model to obtain accurate 3D head models, used for simulationin the context of three-dimensional audio rendering. The second approach presentsa method to fill holes in 3D models, with the use of registered images depicting apattern projected on the real object.Finally, some useful indications about the possible future work in all the presentedfields are given, in order to delineate the developments of this promising direction ofresearch.Item Reconstruction and Analysis of Shapes from 3D Scans(Haar, 2009) Haar, Frank B. terIn this thesis we use 3D laser range scans for the acquisition, reconstruction, and analysis of 3D shapes. 3D laser range scanning has proven to be a fast and effective way to capture the surface of an object in a computer. Thousands of depth measurements represent a part of the surface geometry as a cloud of 3D points and geometric algorithms have been developed to turn such 3D point sets into manageable shapes and shape representations for end users or other algorithms. We use 3D laser range scans to evaluate acquisition and reconstruction systems and algorithms, to fully automate the object reconstruction, to find discriminative face features, for automatic landmark extraction, for face identification with and without expressions, and for the statistical modeling of faces.</p>Item Capturing and Reconstructing the Appearance of Complex 3D Scenes(Fuchs, Christian, 2009-05-29) Fuchs, ChristianIn this thesis, we present our research on new acquisition methods forreflectance properties of real-world objects. Specifically, we firstshow a method for acquiring spatially varying densities in volumes oftranslucent, gaseous material with just a single image. This makes themethod applicable to constantly changing phenomena like smoke withoutthe use of high-speed camera equipment.Furthermore, we investigated how two well known techniques --synthetic aperture confocal imaging and algorithmic descattering --can be combined to help looking through a translucent medium like fogor murky water. We show that the depth at which we can still see anobject embedded in the scattering medium is increased. In a relatedpublication, we show how polarization and descattering based onphase-shifting can be combined for efficient 3D~scanning oftranslucent objects. Normally, subsurface scattering hinders the rangeestimation by offsetting the peak intensity beneath the surface awayfrom the point of incidence. With our method, the subsurfacescattering is reduced to a minimum and therefore reliable 3D~scanningis made possible.Finally, we present a system which recovers surface geometry,reflectance properties of opaque objects, and prevailing lightingconditions at the time of image capture from just a small number ofinput photographs. While there exist previous approaches to recoverreflectance properties, our system is the first to work on imagestaken under almost arbitrary, changing lighting conditions. Thisenables us to use images we took from a community photo collectionwebsite.Item Inverse Tone Mapping(Banterle, 2009-06-04) Banterle, FrancescoThe introduction of High Dynamic Range Imaging in computer graphics has produced a novelty in Imaging that can be compared to the introduction of colour photography or even more. Light can now be captured, stored, processed, and finally visualised without losing information. Moreover, new applications that can exploit physical values of the light have been introduced such as re-lighting of synthetic/real objects, or enhanced visualisation of scenes. However, these new processing and visualisation techniques cannot be applied to movies and pictures that have been produced by photography and cinematography in more than one hundred years. This thesis introduces a general framework for expanding legacy content into High Dynamic Range content. The expansion is achieved avoiding artefacts, producing images suitable for visualisation and re-lighting of synthetic/real objects. Moreover, it is presented a methodology based on psychophysical experiments and computational metrics to measure performances of expansion algorithms. Finally, a compression scheme, inspired by the framework, for High Dynamic Range Textures, is proposed and evaluated.Item Feature Extraction for Visual Analysis of DW-MRI Data(Schultz, Thomas, 2009-06-18) Schultz, ThomasDiffusion Weighted Magnetic Resonance Imaging (DW-MRI) is a recentmodality to investigate the major neuronal pathways of the humanbrain. However, the rich DW-MRI datasets cannot be interpreted withoutproper preprocessing. In order to achieve understandablevisualizations, this dissertation reduces the complex data to relevantfeatures.<br>The first part is inspired by topological features in flow data. Novelfeatures reconstruct fuzzy fiber bundle geometry from probabilistictractography results. The topological properties of existing featuresthat extract the skeleton of white matter tracts are clarified, andthe core of regions with planar diffusion is visualized.<br>The second part builds on methods from computer vision. Relevantboundaries in the data are identified via regularized eigenvaluederivatives, and boundary information is used to segment anisotropyisosurfaces into meaningful regions. A higher-order structure tensoris shown to be an accurate descriptor of local structure in diffusiondata.<br>The third part is concerned with fiber tracking. Streamlinevisualizations are improved by adding features from structural MRI ina way that emphasizes the relation between the two types of data, andthe accuracy of streamlines in high angular resolution data isincreased by modeling the estimation of crossing fiber bundles as alow-rank tensor approximation problem.Item Filtering and Optimization Strategies for Markerless Human Motion Capture with Skeleton-Based Shape Models(Gall, Juergen, 2009-07-07) Gall, JuergenSince more than 2000 years, people have been interested in understanding and analyzing themovements of animals and humans which lead to the development of advanced computer systemsfor motion capture. Although marker-based systems for motion analysis are commerciallysuccessful, capturing the performance of a human or an animal from a multi-view video sequencewithout the need for markers is still a challenging task. The most popular methods formarkerless human motion capture are model-based approaches that rely on a surface model ofthe human with an underlying skeleton. In this context, markerless motion capture seeks for thepose, i.e., the position, orientation, and configuration of the human skeleton that is best explainedby the image data. In order to address this problem, we discuss the two questions:1. What are good cues for human motion capture? Typical cues for motion capture are silhouettes,edges, color, motion, and texture. In general, a multi-cue integration is necessary fortracking complex objects like humans since all these cues come along with inherent drawbacks.Besides the selection of the cues to be combined, reasonable information fusion is a commonchallenge in many computer vision tasks. Ideally, the impact of a cue should be large in situationswhen its extraction is reliable, and small, if the information is likely to be erroneous. Tothis end, we propose an adaptive weighting scheme that combines complementary cues, namelysilhouettes on one side and optical flow as well as local descriptors on the other side. Whereassilhouette extraction works best in case of homogeneous objects, optical flow computation andlocal descriptors perform better on sufficiently structured objects. Besides image-based cues, wealso propose a statistical prior on anatomical constraints that is independent of motion patterns.Relying only on image features that are tracked over time does not prevent the accumulationof small errors which results in a drift away from the target object. The error accumulationbecomes even more problematic in the case of multiple moving objects due to occlusions. Tosolve the drift problem for tracking, we propose an analysis-by-synthesis framework that usesreference images to correct the pose. It comprises an occlusion handling and is successfullyapplied to crash test video analysis.2. Is human motion capture a filtering or an optimization problem? Model-based human motioncapture can be regarded as a filtering or an optimization problem. While local optimizationoffers accurate estimates but often looses track due to local optima, particle filtering can recoverfrom errors at the expense of a poor accuracy due to overestimation of noise. In order to overcomethe drawbacks of local optimization, we introduce a novel global stochastic optimizationapproach for markerless human motion capturing that is derived from the mathematical theoryon interacting particle systems. We call the method interacting simulated annealing (ISA) sinceit is based on an interacting particle system that converges to the global optimum similar tosimulated annealing. It estimates the human pose without initial information, which is a challengingoptimization problem in a high dimensional space. Furthermore, we propose a trackingframework that is based on this optimization technique to achieve both the robustness of filteringstrategies and a remarkable accuracy.In order to benefit from optimization and filtering, we introduce a multi-layer framework thatcombines stochastic optimization, filtering, and local optimization. While the first layer relieson interacting simulated annealing, the second layer refines the estimates by filtering and localoptimization such that the accuracy is increased and ambiguities are resolved over time withoutimposing restrictions on the dynamics.In addition, we propose a system that recovers not only the movement of the skeleton, but alsothe possibly non-rigid temporal deformation of the 3D surface. While large scale deformationsor fast movements are captured by the skeleton pose and approximate surface skinning, truesmall scale deformations or non-rigid garment motion are captured by fitting the surface to thesilhouette. In order to make automatic processing of large data sets feasible, the skeleton-basedpose estimation is split into a local one and a lower dimensional global one by exploiting thetree structure of the skeleton.Our experiments comprise a large variety of sequences for qualitative and quantitative evaluationof the proposed methods, including a comparison of global stochastic optimization withseveral other optimization and particle filtering approaches.Item High Quality Dynamic Reflectance and Surface Reconstruction from Video(Ahmed, Naveed, 2009-07-10) Ahmed, NaveedThe creation of high quality animations of real-world human actors has long been a challenging problem in computer graphics. It involves the modeling of the shape of the virtual actors, creating their motion, and the reproduction of very fine dynamic details. In order to render the actor under arbitrary lighting, it is required that reflectance properties are modeled for each point on the surface. These steps, that are usually performed manually by professional modelers, are time consuming and cumbersome.<br> In this thesis, we show that algorithmic solutions for some of the problems that arise in the creation of high quality animation of real-world people are possible using multi-view video data. First, we present a novel spatio-temporal approach to create a personalized avatar from multi-view video data of a moving person. Thereafter, we propose two enhancements to a method that captures human shape, motion and reflectance properties of a moving human using eight multi-view video streams. Afterwards we extend this work, and in order to add very fine dynamic details to the geometric models, such as wrinkles and folds in the clothing, we make use of the multi-view video recordings and present a statistical method that can passively capture the fine-grain details of time-varying scene geometry. Finally, in order to reconstruct structured shape and animation of the subject from video, we present a dense 3D correspondence finding method that enables spatiotemporally coherent reconstruction of surface animations directly from multi-view video data.<br>These algorithmic solutions can be combined to constitute a complete animation pipeline for acquisition, reconstruction and rendering of high quality virtual actors from multi-view video data. They can also be used individually in a system that requires the solution of a specific algorithmic sub-problem. The results demonstrate that using multi-view video data it is possible to find the model description that enables realistic appearance of animated virtual actors under different lighting conditions and exhibits high quality dynamic details in the geometry.Item Expressive Visualization and Rapid Interpretation of Seismic Volumes(Patel, 2009-08-01) Patel, DanielOne of the most important resources in the world today is energy. Oil and gas provide two thirds of the world energy consumption, making the world completely dependent on it. Locating and recovering the remaining oil and gas reserves will be of high focus in society until competitive energy sources are found. The search for hydrocarbons is broadly speaking the topic of this thesis. Seismic measurements of the subsurface are collected to discover oil and gas trapped in the ground. Identifying oil and gas in the seismic measurements requires visualization and interpretation. Visualization is needed to present the data for further analysis. Interpretation is performed to identify important structures. Visualization is again required for presenting these structures to the user. This thesis investigates how computer assistance in producing high-quality visualizations and in interpretation can result in expressive visualization and rapid interpretation of seismic volumes. Expressive visualizations represent the seismic data in an easily digestible, intuitive and pedagogic form. This enables rapid interpretation which accelerates the nding of important structures. 3Item Robust and Efficient Processing Techniques for Staticand Dynamic Geometric Data(Schall, Oliver, 2009-08-28) Schall, OliverGenerating high quality geometric representations from real-world objects is a fundamental problem in computer graphics which is motivated bymanifold applications. They comprise image synthesis for movie productionor computer games but also industrial applications such as quality assurance in mechanical engineering, the preservation of cultural heritage and the medical adaptation of prostheses or orthoses. Common demands of these applications on their underlying algorithms are robustness and efficiency. In addition, technological improvements of scanning devices and cameras which allow for the acquisition of new data types such as dynamic geometric data, create novel requirements which rise new challenges for processing algorithms.This dissertation focuses on these aspects and presents differentcontributions for flexible, efficient and robust processing of staticand time-varying geometric data. Two techniques focus on the problemof denoising. A statistical filtering algorithm for point cloud databuilding on non-parametric density estimation is introduced as well asa neighborhood filter for static and time-varying range data which isbased on a novel non-local similarity measure. The third contributionunifies partition of unity decomposition and a global surfacereconstruction algorithm based on the Fast Fourier Transformwhich results in a novel, robust and efficient reconstructiontechnique. Concluding, two flexible and versatile tools for designingscalar fields on meshes are presented which are useful to facilitate acontrollable quadrangular remeshing.Item Audio and Visual Rendering with Perceptual Foundations(Bonneel, 2009-09-15) Bonneel, NicolasRealistic visual and audio rendering still remains a technical challenge. Indeed, typicalcomputers do not cope with the increasing complexity of today's virtual environments, bothfor audio and visuals, and the graphic design of such scenes require talented artists.In the first part of this thesis, we focus on audiovisual rendering algorithms for complexvirtual environments which we improve using human perception of combined audioand visual cues. In particular, we developed a full perceptual audiovisual rendering engineintegrating an efficient impact sounds rendering improved by using our perception ofaudiovisual simultaneity, a way to cluster sound sources using human's spatial tolerancebetween a sound and its visual representation, and a combined level of detail mechanismfor both audio and visuals varying the impact sounds quality and the visually rendered materialquality of the objects. All our crossmodal effects were supported by the prior workin neuroscience and demonstrated using our own experiments in virtual environments.In a second part, we use information present in photographs in order to guide a visualrendering. We thus provide two different tools to assist casual artists such as gamers, orengineers. The first extracts the visual hair appearance from a photograph thus allowingthe rapid customization of avatars in virtual environments. The second allows for a fastpreviewing of 3D scenes reproducing the appearance of an input photograph following auser's 3D sketch.We thus propose a first step toward crossmodal audiovisual rendering algorithms anddevelop practical tools for non expert users to create virtual worlds using photograph'sappearance.Item Template based shape processing(Stoll, Carsten, 2009-09-30) Stoll, CarstenAs computers can only represent and process discrete data, informationgathered from the real world always has to be sampled. While it isnowadays possible to sample many signals accurately and thus generatehigh-quality reconstructions (for example of images and audio data),accurately and densely sampling 3D geometry is still a challenge. Thesignal samples may be corrupted by noise and outliers, and contain largeholes due to occlusions. These issues become even more pronounced whenalso considering the temporal domain. Because of this, developing methodsfor accurate reconstruction of shapes from a sparse set of discrete datais an important aspect of the computer graphics processing pipeline. In this thesis we propose novel approaches to including semantic knowledgeinto reconstruction processes using template based shape processing. Weformulate shape reconstruction as a deformable template fitting process,where we try to fit a given template model to the sampled data. Thisapproach allows us to present novel solutions to several fundamentalproblems in the area of shape reconstruction. We address static problemslike constrained texture mapping and semantically meaningful hole-fillingin surface reconstruction from 3D scans, temporal problems such as meshbased performance capture, and finally dynamic problems like theestimation of physically based material parameters of animated templates.Item Manipulations d'image expressives pour une variété de représentations visuelles - Expressive Image Manipulations for a Variety of Visual Representations(Bousseau, 2009-10-15) Bousseau, AdrienLa communication visuelle tire profit de la grande variété d'apparences qu'une image peut avoir. En ignorant les détails, les images simplifiées concentrent l'attention de l'observateur sur le contenu essentiel à transmettre. Les images stylisées, qui diffèrent de la réalité, peuvent suggérer une information subjective ou imaginaire. Des variations plus subtiles, comme le changement de l'éclairage dans une photographie, ont également un impact direct sur la façon dont le message transmis va être interprété. Le but de cette thèse est de permettre à un utilisateur de manipuler le contenu visuel et créer des images qui correspondent au message qu'il cherche à transmettre. Nous proposons plusieurs manipulations qui modifient, simplifient ou stylisent des images pour augmenter leur pouvoir d'expression. Nous présentons d'abord deux méthodes pour enlever les détails d'une photographie ou d'une vidéo. Le résultat de cette simplification met en valeur les structures importantes de l'image. Nous introduisons ensuite une nouvelle primitive vectorielle, nommée Courbe de Diffusion, qui facilite la création de dégradés de couleurs et de flou dans des images vectorielles. Les images créées avec des courbes de diffusion présentent des effets complexes qui sont difficiles à reproduire avec les outils vectoriels existants. Dans une seconde partie, nous proposons deux algorithmes pour la création d'animations stylisées à partir de vidéos et de scènes 3D. Ces deux méthodes produisent des animations qui ont l'apparence 2D de média traditionnels comme l'aquarelle. Nous décrivons enfin une approche pour décomposer l'information d'illumination et de réflectance dans une photographie. Nous utilisons des indications utilisateurs pour résoudre ce problème sous-contraint. Les différentes manipulations d'image proposées dans ce mémoire facilitent la création d'une variété de représentations visuelles, comme illustré par nos résultats. - Visual communication greatly benefits from the large variety of appearances that an image can take. By neglecting spurious details, simplified images focus the attention of an observer on the essential message to transmit. Stylized images, that depart from reality, can suggest subjective or imaginary information. More subtle variations, such as change of lighting in a photograph can also have a dramatic effect on the interpretation of the transmitted message. The goal of this thesis is to allow users to manipulate visual content and create images that corresponds to their communication intent. We propose a number of manipulations that modify, simplify or stylize images in order to improve their expressive power. We first present two methods to remove details in photographs and videos. The resulting simplification enhances the relevant structures of an image. We then introduce a novel vector primitive, called Diffusion Curves, that facilitates the creation of smooth color gradients and blur in vector graphics. The images created with diffusion curves contain complex image features that are hard to obtain with existing vector primitives. In the second part of this manuscript we propose two algorithms for the creation of stylized animations from 3D scenes and videos. The two methods produce animations with the 2D appearance of traditional media such as watercolor. Finally, we describe an approach to decompose the illumination and reflectance components of a photograph. We make this ill-posed problem tractable by propagating sparse user indications. This decomposition allows users to modify lighting or material in the depicted scene. The various image manipulations proposed in this dissertation facilitates the creation of a variety of visual representations, as illustrated by our results.Item SELF-DELAUNAY MESHES FOR SURFACES(Dyer, 2010) Dyer, RamsayIn the Euclidean plane, a Delaunay triangulation can be characterized by the requirementthat the circumcircle of each triangle be empty of vertices of all other triangles. For triangulatinga surface S in R3, the Delaunay paradigm has typically been employed in the formof the restricted Delaunay triangulation, where the empty circumcircle property is definedby using the Euclidean metric in R3 to measure distances on the surface. More recently, theintrinsic (geodesic) metric of S has also been employed to define the Delaunay condition.In either case the resulting mesh M is known to approximate S with increasing accuracyas the density of the sample points increases. However, the use of the reference surface Sto define the Delaunay criterion is a serious limitation. In particular, in the absence of theoriginal reference surface, there is no way of verifying if a given mesh meets the criterion.We define a self-Delaunay mesh as a triangle mesh that is a Delaunay triangulation ofits vertex set with respect to the intrinsic metric of the mesh itself. This yields a discretesurface representation criterion that can be validated by the properties of the mesh alone,independent of any reference surface the mesh is supposed to represent. The intrinsic Delaunaytriangulation that characterizes self-Delaunay meshes makes them a natural domainfor discrete differential geometry, and the discrete exterior calculus in particular.We examine self-Delaunay meshes and their relationship with other Delaunay structuresfor surface representation. We study sampling conditions relevant to the intrinsic approach,and compare these with traditional sampling conditions which are based on extrinsic quantitiesand distances in the ambient Euclidean space. We also provide practical and provablycorrect algorithms for constructing self-Delaunay meshes. Of particular interest in thiscontext is the extrinsic edge flipping algorithm which extends the familiar algorithm forproducing planar Delaunay triangulations.Item Perceptually-motivated, Interactive Rendering and Editing of Global Illumination(Universitaet des Saarlands, 22.12.2009) Ritschel, TobiasThis thesis proposes several new perceptually-motivated techniques to synthesize, edit and enhance depiction of three-dimensional virtual scenes. Finding algorithmsthat fit the perceptually economic middle ground between artistic depiction and full physical simulation is the challenge taken in this work. First, we will present threeinteractive global illumination rendering approaches that are inspired by perception to efficiently depict important light transport. Those methods have in commonto compute global illumination in large and fully dynamic scenes allowing for light, geometry, and material changes at interactive or real-time rates. Further,this thesis proposes a tool to edit reflections, that allows to bend physical laws to match artistic goals by exploiting perception. Finally, this work contributes apost-processing operator that depicts high contrast scenes in the same way as artists do, by simulating itItem Applications of temporal coherence in real-time rendering(Scherzer, 27.11.2009) Scherzer, DanielReal-time rendering imposes the challenging task of creating a new rendering of an input scene at least 60 times a second. Although computer graphics hardware has made staggering advances in terms of speed and freedom of programmability, there still exist a number of algorithms that are too expensive to be calculated in this time budget, like exact shadows or an exact global illumination solution. One way to circumvent this hard time limit is to capitalize on temporal coherence to formulate algorithms incremental in time. The main thesis of this work is that temporal coherence is a characteristic of real-time graphics that can be used to redesign well-known rendering methods to become faster, while exhibiting better visual fidelity. To this end we present our adaptations of algorithms from the fields of exact hard shadows, physically correct soft shadows and fast discrete LOD blending, in which we have successfully incorporated temporal coherence. Additionally, we provide a detailed context of previous work not only in the field of temporal coherence, but also in the respective fields of the presented algorithms.Item On Visualization and Reconstruction from Non-uniform Point Sets(Vuçini, 27.11.2009) Vuçini, EraldTechnological and research advances in both acquisition and simulation devices provide continuously increasing high-resolution volumetric data that by far exceed today's graphical and display capabilities. Non-uniform representations offer a way of balancing this deluge of data by adaptively measuring (sampling) according to the importance (variance) of the data. Also, in many real-life situations the data are known only on a non-uniform representation. Processing of non-uniform data is a non-trivial task and hence more difficult when compared to processing of regular data. Transforming from non-uniform to uniform representations is a well-accepted paradigm in the signal processing community. In this thesis we advocate such a concept. The main motivation for adopting this paradigm is that most of the techniques and methods related to signal processing, data mining and data exploration are well-defined and stable for Cartesian data, but generally are non-trivial to apply to non-uniform data. Among other things, this will allow us to better exploit the capabilities of modern GPUs.In non-uniform representations sampling rates can vary drastically even by several orders of magnitude, making the decision on a target resolution a non-trivial trade-off between accuracy and efficiency. In several cases the points are spread non-uniformly with similar density across the volume, while in other cases the points have an enormous variance in distribution. In this thesis we present solutions to both cases. For the first case we suggest computing reconstructions of the same volume in different resolutions based on the level of detail we are interested in. The second case scenario is the main motivation for proposing a multi-resolution scheme, where the scale of reconstruction is decided adaptively based on the number of points in each subregion of the whole volume.We introduce a novel framework for 3D reconstruction and visualization from non-uniform scalar and vector data. We adopt a variational reconstruction approach. In this method non-uniform point sets are transformed to a uniform representation consisting of B-spline coefficients that are attached to the grid. With these coefficients we can define a C2 continuous function across the whole volume. Several testings were performed in order to analyze and fine-tune our framework. All the testings and the results of this thesis offer a view from a new and different perspective to the visualization and reconstruction from non-uniform point sets.Item GPU-based Multi-Volume Rendering of Complex Data in Neuroscience and Neurosurgery(Beyer, 27.11.2009) Beyer, JohannaRecent advances in image acquisition technology and its availability in the medical and bio-medical fields have lead to an unprecedented amount of high-resolution imaging data. However, the inherent complexity of this data, caused by its tremendous size, complex structure or multi-modality poses several challenges for current visualization tools. Recent developments in graphics hardware architecture have increased the versatility and processing power of today's GPUs to the point where GPUs can be considered parallel scientific computing devices. The work in this thesis builds on the current progress in image acquisition techniques and graphics hardware architecture to develop novel 3D visualization methods for the fields of neurosurgery and neuroscience. The first part of this thesis presents an application and framework for planning of neurosurgical interventions. Concurrent GPU-based multi-volume rendering is used to visualize multiple radiological imaging modalities, delineating the patient's anatomy, neurological function, and metabolic processes. Additionally, novel interaction metaphors are introduced, allowing the surgeon to plan and simulate the surgial approach to the brain based on the individual patient anatomy. The second part of this thesis focuses on GPU-based volume rendering techniques for large and complex EM data, as required in the field of neuroscience. A new mixed-resolution volume ray-casting approach is presented, which circumvents artifacts at block boundaries of different resolutions. NeuroTrace is introduced, an application for interactive segmentation and visualization of neural processes in EM data. EM data is extremely dense, heavily textured and exhibits a complex structure of interconnected nerve cells, making it difficult to achieve high-quality volume renderings. Therefore, this thesis presents a novel on-demand nonlinear noise removal and edge detection method which allows to enhance important structures (e.g., myelinated axons) while de-emphasizing less important regions of the data. In addition to the methods and concepts described above, this thesis tries to bridge the gap between state-of-the-art visualization research and the use of those visualization methods in actual medical and bio-medical applications.Item Analysis and Visualization of Industrial CT Data(Heinzl, Dec 2008) Heinzl, ChristophDie industrielle 3D Röntgencomputertomographie (3DCT) steht derzeitan der Schwelle von einer zerstörungsfreien Werkstoffprüfmethode hin zueiner genormten Methode für dimensionales Messen. 3DCT wird vor allemim Bereich der Erstmusterprüfung von neuen Komponenten eingesetzt,um die Nachteile und Einschränkungen bisheriger Methoden zu überwinden.Eine steigende Anzahl von Firmen vertraut daher auf 3DCT undsporadisch wird 3DCT bereits von einigen Pionieren für die Qualitätskontrollein der Produktion eingesetzt. Dennoch ist die 3DCT eine sehr jungeMethode mit einigen Nachteilen, die großen Einfluss auf das Messergebnishaben. Einige der größten Nachteile von 3DCT im Bereich der Qualitätssicherungsind:Artefakte ändern die Grauwerte im Datensatz und generieren künstlicheStrukturen, die in Realität nicht vorhanden sind.Diskretisierung bewirkt Unregelmäßigkeiten in den Grauwertenentsprechend des Abtasttheorems von Nyquist-Shannon.Informationen bezüglich Unsicherheit der Daten gehen bei der Extraktionvon dimensionalen Messmerkmalen verloren.Spezifikationen and Einschränkungen der einzelnen Komponenten undder Bauweise des 3DCTs limitieren die erreichbare Messgenauigkeit.Diese Dissertation trägt zum Stand der Technik durch algorithmische Lösungenvon typischen industriellen Problemen im Bereich der Metrologiemittels 3DCT bei. Das Hauptaugenmerk der präsentierten Arbeit liegtin der Entwicklung und Implementierung von neuen Prozessketten, diefür den täglichen industriellen Einsatz im Bereich der Qualitätssicherungoptimiert sind. Geeignete, einfach verständliche Visualisierungsmethodenwerden evaluiert und angewendet, um einen Einblick in die generiertenMessdaten zu ermöglichen. Im Speziellen werden drei Prozesskettenpräsentiert, die einige der wesentlichen Aspekte der Metrologie mittels3DCT abdecken. Die betrachteten Aspekte sind robuste Oberflächeextraktion,Artefaktreduzierung mittels Dual Energy CT, lokale Oberflächeextraktionvon Multimaterialkomponenten und statistische Analyse vonMultimaterialkomponenten. Die generierten Ergebnisse jeder Prozesskettewerden anhand von Testteilen und typischen Industriebauteilen demonstriertund verifiziert - Industrial X-Ray 3D computed tomography (3DCT) is on the edge ofadvancing from a non destructive testing method to a fully standardizedmeans of dimensional measurement for every day industrial use. Currently3DCT has drawn attention especially in the area of first part inspections ofnew components, mainly in order to overcome limitations and drawbacksof common methods. Yet an increasing number of companies is benefittingfrom industrial 3DCT and sporadically the first pioneers start using industrial3DCT for quality control in the production phase of a component. As3DCT is still a very young technology of industrial quality control, thismethod also faces severe problems, which seriously affect measurementresults. Some of the major drawbacks for quality control are the following:Artefacts modify the spatial greyvalues, generating artificial structures inthe datasets, which do not correspond to reality.Discrete sampling introduces further irregularities due to the Nyquist-Shannon sampling theorem.Uncertainty information is missing when extracting dimensional measurementfeatures.Specifications and limitations of the components and the special setup a3DCT constrain the best achievable measurement precision.This thesis contributes to the state of the art by algorithmic evaluationof typical industrial tasks in the area of dimensional measurement using3DCT. The main focus lies in the development and implementation of novelpipelines for everyday industrial use including comparisons to commonmethods. Convenient and easy to understand means of visualization are evaluated and used to provide insight into the generated results. In particularthree pipelines are introduced, which cover some of the major aspectsconcerning metrology using industrial 3DCT. The considered aspectsare robust surface extraction, artefact reduction via dual energy CT, localsurface extraction of multi-material components, and statistical analysisof multi-material components. The generated results of each pipeline aredemonstrated and verified using test specimens as well as real world components.Item Comprehensive Visualization of Cardiac MRI Data(Termeer, Dec 2008) Termeer, Maurice AlainKoronare Herzkrankheit ist eine der führenden Todesursachen in derwestlichen Welt. Die kontinuierliche Verbesserung der Magnetresonanztomographie(MRT) erleichtert genauere Diagnosen, indem sie immerdetailliertere Informationen über die Lebensfähigkeit, das Funktionieren,die Durchblutung und die Anatomie des Herzens eines Patienten liefert.Diese zunehmende Menge an Informationen schafft die Notwendigkeit füreffizientere und effektivere Mittel der Verarbeitung dieser Daten.Diese Dissertation präsentiert mehrere neue Techniken, die eine umfassendereVisualisierung des Patienten bei der Diagnose von Erkrankungender Herzkranzgefäße mittels MRT unterstützen. Das volumetrische Polardiagramwird als Erweiterung des Polardiagrams, welches eine bestehendeVisualusierungstechnik in der klinischen Praxis ist, eingeführt. Diesesneuartige Konzept bietet eine umfassendere Sicht auf die Lebensfähigkeitdes Herzens eines Patienten, indem detaillierte Informationen über dieTransmuralität der Narbe ohne Diskontinuitäten bereitgestellt werden.Anatomische Zusammenhänge gehen in abstrakten Darstellungen vonDaten häufig verloren. Darüberhinaus liefern einige Arten von Scansrelativ wenig anatomischen Kontext. Mehrere Techniken zur Wiederherstellungdes anatomischen Bezugs werden vorgestellt. Die primärenKoronararterien sind in einem Scan des ganzen Herzens segmentiert undwerden auf ein volumetrisches Polardiagram abgebildet. Hierbei wirdder abstrakten Repräsentation anatomischer Kontext hinzugefügt. Ebenso,werden segmentierte späte Anreicherungs Daten zusammen mit einerdrei-dimensionalen Segmentierung des patientenspezifischen Herzmuskelsund koronarer Anatomie dargestellt. Darüberhinaus werden koronareVersorgungsgebiete aus den patientenspezifischen Daten berechnet. Diesbedeutet eine Verbesserung gegenüber Modellen welche auf Bevölkerungsdurchschnittenbasieren.Informationen über die Durchblutung des Herzmuskels welche ausMRT-Aufnahmen abgeleitet werden können sind in der Regel von relativgeringer Auflösung. Unter Verwendung hochauflösender anatomischenDaten wird ein Konzept für die Visualisierung simulierter Durchblutungdes Herzmuskels präsentiert. Dabei wird die detaillierte Information überdie Durchblutung genutzt. Schließlich, wird eine wirklich umfassendeVisualisierung einer Herz-MRT-Untersuchung erforscht. Dabei werden Scans des ganzen Herzens, der Herzvitalität, der Herzfunktion und derDurchblutung in einer einzigen Visualisierung kombiniert. Die eingeführtenKonzepte fördern den Aufbau eines umfassenderen Überblicksüber den Patienten. Die dabei zusätzlich gewonnene Information kann fürden Diagnoseprozess von Nutzem sein. - Coronary artery disease is one of the leading causes of death in the westernworld. The continuous improvements in magnetic resonance imagingtechnology facilitate more accurate diagnoses by providing increasinglymore detailed information on the viability, functioning, perfusion, andanatomy of a patient s heart. This increasing amount of information createsthe need for more efficient and more effective means of processing thesedata.This thesis presents several novel techniques that facilitate a morecomprehensive visualization of a patient s heart to assist in the diagnosisof coronary artery disease using magnetic resonance imaging (MRI). Thevolumetric bull s eye plot is introduced as an extension of an existingvisualization technique used in clinical practice the bull s eye plot. Thisnovel concept offers a more comprehensive view on the viability of apatient s heart by providing detailed information on the transmurality ofscar while not suffering from discontinuities.Anatomical context is often lost due to abstract representations of data,or may be scarce due to the nature of the scanning protocol. Severaltechniques to restore the relation to anatomy are presented. The primarycoronary arteries are segmented in a whole heart scan and mapped ontoa volumetric bull s eye plot, adding anatomical context to an abstractrepresentation. Similarly, segmented late enhancement data are renderedalong with a three-dimensional segmentation of the patient-specific myocardialand coronary anatomy. Additionally, coronary supply territoriesare computed from patient-specific data as an improvement over modelsbased on population averages.Information on the perfusion of the myocardium provided by MRI istypically of fairly low resolution. Using high-resolution anatomical data,an approach to visualize simulated myocardial perfusion is presented,taking full advantage of the detailed information on perfusion. Finally, atruly comprehensive visualization of a cardiac MRI exam is explored bycombining whole heart, late enhancement, functional, and perfusion scansin a single visualization. The concepts introduced help to build a morecomprehensive view of the patient and the additional information mayprove to be beneficial for the diagnostic process.