Computer Graphics is about digital models for threedimensional geometric objects as well as images. These shapes and images may represent approximations of the real world or could be synthetic, i.e., exist only in the computer. Goals of computer graphics research are the generation of plausible and informative images, and computation with reasonable resources, i.e. in a short amount of time with little storage requirements. The models and algorithms for this task combine knowledge from different areas of mathematics and computer science.
ABC: A Big CAD Model Dataset For Geometric Deep Learning (CVPR 2019)
We introduce ABC-Dataset, a collection of one million Computer-Aided Design (CAD) models for research of geometric deep learning methods and applications. Each model is a collection of explicitly parametrized curves and surfaces, providing ground truth for differential quantities, patch segmentation, geometric feature detection, and shape reconstruction. Sampling the parametric descriptions of surfaces and curves allows generating data in different formats and resolutions, enabling fair comparisons for a wide range of geometric learning algorithms. As a use case for our dataset, we perform a large-scale benchmark for estimation of surface normals, comparing existing data driven methods and evaluating their performance against both the ground truth and traditional normal estimation methods.
CHI’19: I Can See What You Think: The Mental Image Revealed by Gaze Tracking
Humans involuntarily move their eyes when retrieving an image from memory. This motion is often similar to actually observing the image. We suggest to exploit this behavior as a new modality in human computer interaction, using the motion of the eyes as a descriptor of the image. Interaction requires the user’s eyes to be tracked but no voluntary physical activity. We perform a controlled experiment and develop matching techniques using machine learning to investigate if images can be discriminated based on the gaze patterns recorded while users merely think about image. Our results indicate that image retrieval is possible with an accuracy significantly above chance. We also show that this result generalizes to images not used during training of the classifier and extends to uncontrolled settings in a realistic scenario.
Siggraph Asia 2018: Tracking the Gaze on Objects in 3D: How do People Really Look at the Bunny?
We provide the first large dataset of human fixations on physical 3D objects presented in varying viewing conditions and made of different materials. Our experimental setup is carefully designed to allow for accurate calibration and measurement. We estimate a mapping from the pair of pupil positions to 3D coordinates in space and register the presented shape with the eye tracking setup. By modeling the fixated positions on 3D shapes as a probability distribution, we analysis the similarities among different conditions. The resulting data indicates that salient features depend on the viewing direction. Stable features across different viewing directions seem to be connected to semantically meaningful parts. We also show that it is possible to estimate the gaze density maps from view dependent data. The dataset provides the necessary ground truth data for computational models of human perception in 3D.
Siggraph Asia 2018: Factor Once: Reusing Cholesky Factorizations on Sub-Meshes
A common operation in geometry processing is solving symmetric and positive semi-definite systems on a subset of a mesh with conditions for the vertices at the boundary of the region. This is commonly done by setting up the linear system for the sub-mesh, factorizing the system (potentially applying preordering to improve sparseness of the factors), and then solving by back-substitution. This approach suffers from a comparably high setup cost for each local operation. We propose to reuse factorizations defined on the full mesh to solve linear problems on sub-meshes. We show how an update on sparse matrices can be performed in a particularly efficient way to obtain the factorization of the operator on a sun-mesh significantly outperforming general factor updates and complete refactorization. We analyze the resulting speedup for a variety of situations and demonstrate that our method outperforms factorization of a new matrix by a factor of up to 10 while never being slower in our experiments.
Prof. Alexa selected as Editor-in-Chief of ACM Transactions on Graphics
Prof. Alexa has been selected as Editor in Chief of ACM Transactions on Graphics (TOG), the leading technical journal in the field of computer graphics.
Prof. Alexa elected as Fellow of Eurographics
Each year, the European Association for Computer Graphics elects up to three members for their longstanding contributions to be Fellows of the Association. Prof. Alexa has been elected as one of two new Fellows in 2018. Citation and more information.
Siggraph Asia 2017: Localized solutions of sparse linear systems for geometry processing
Computing solutions to linear systems is a fundamental building block of many geometry processing algorithms. In many cases the Cholesky factorization of the system matrix is computed to subsequently solve the system, possibly for many right-hand sides, using forward and back substitution. We demonstrate how to exploit sparsity in both the right-hand side and the set of desired solution values to obtain significant speedups. The method is easy to implement and potentially useful in any scenarios where linear problems have to be solved locally. We show that this technique is useful for geometry processing operations, in particular we consider the solution of diffusion problems. All problems profit significantly from sparse computations in terms of runtime, which we demonstrate by providing timings for a set of numerical experiments.
UIST 2017: HeatSpace – Automatic Placement of Displays by Empirical Analysis of User Behavior
We present HeatSpace, a system that records and empirically analyzes user behavior in a space and automatically suggests positions and sizes for new displays. The system uses depth cameras to capture 3D geometry and users’ perspectives over time. To derive possible display placements, it calculates volumetric heatmaps describing geometric persistence and planarity of structures inside the space. It evaluates visibility of display poses by calculating a volumetric heatmap describing occlusions, position within users’ field of view, and viewing angle. Optimal display size is calculated through a heatmap of average viewing distance. Based on the heatmaps and user constraints we sample the space of valid display placements and jointly optimize their positions. This can be useful when installing displays in multi-display environments such as meeting rooms, offices, and train stations.
SMI 2017: Unsharp Masking Geometry Improves 3D Prints
Mass market digital manufacturing devices are severely limited in accuracy and material, resulting in a significant gap between the appearance of the virtual and the real shape. In imaging as well as rendering of shapes, it is common to enhance features so that they are more apparent. We provide an approach for feature enhancement that directly operates on the geometry of a given shape, with particular focus on improving the visual appearance for 3D printing. The technique is based on unsharp masking, modified to handle arbitrary free-form geometry in a stable, efficient way, without causing large scale deformation. On a series of manufactured shapes we show how features are lost as size of the object decreases, and how our technique can compensate for this. We evaluate this effect in a human subject experiment and find significant preference for modified geometry.
CHI 2017: Changing the Appearance of Real-World Objects by Modifying Their Surroundings
We present an approach to alter the perceived appearance of physical objects by controlling their surrounding space. Many real-world objects cannot easily be equipped with displays or actuators in order to change their shape. While common approaches such as projection mapping enable changing the appearance of objects without modifying them, certain surface properties (e. g. highly reflective or transparent surfaces) can make employing these techniques difficult. In this work, we present a conceptual design exploration on how the appearance of an object can be changed by solely altering the space around it, rather than the object itself. In a proof-of-concept implementation, we place objects onto a tabletop display and track them together with users to display perspective-corrected 3D graphics for augmentation. This enables controlling properties such as the perceived size, color, or shape of objects. We characterize the design space of our approach and demonstrate potential applications. For example, we change the contour of a wallet to notify users when their bank account is debited. We envision our approach to gain in importance with increasing ubiquity of display surfaces.