22.12.2022: A state of the art report (STAR) conditionally accepted at EUROGRAPHICS 2023.
16.12.2022: We are going to organise a
Workshop on Quantum Computer Vision and Machine Learning at CVPR 2023; in collaboration with
Tongyang Li, Jan-Nico Zaech, Martin Danelljan, Jacob Biamonte and Tolga Birdal. More details coming soon.
13.12.2022: Two articles conditionally accepted at EUROGRAPHICS 2023.
18.10.2021: An article is accepted at TPAMI (HandVoxNet++). Our method is ranked first on the HANDS19 challenge dataset (Task 1: Depth-Based 3D Hand Pose Estimation) as of August 2020.
08.09.2021: Dennis Willsch gave a 4DQV Lecture Series talk titled Applications on Quantum Annealers at Jülich Supercomputing Centre. The video can be found here.
05.07.2021: The source code of PhysCap is now available here.
03.05.2021: I am starting a research group on 4D and quantum computer vision and accepting now applications for PhD positions, post-doc positions and internships (including theses internships and remote internships).
PhD positions, post-doc positions and internships are available. Check how to apply.
If you are interested in a bachelor/master thesis, an internship, research immersion lab (RIL) or a HiWi position on one of the topics listed below, do not hesitate to reach me out.
Research Profile
I am currently leading "4D and Quantum Vision" group at Max Planck Institute for Informatics, D6 Department. The focus of our team lies on 3D reconstruction and analysis of general deformable scenes, 3D reconstruction of the human body and matching problems on point sets and graphs. We are interested in neural approaches (both supervised and unsupervised), physics-based methods as well as new hardware and sensors (e.g., quantum computers and event cameras).
Many research questions at the intersection of computer graphics, computer vision and machine learning involve challenging search problems (e.g., graph matching) or the optimisation of non-convex objectives. For such problems, we develop new algorithmic formulations that can be solved on modern adiabatic quantum annealers or universal quantum computers and investigate which advantages these approaches offer compared to existing classical methods.
Our research interests include (but are not limited to):
3D Reconstruction and Neural Rendering of Rigid and Non-Rigid Scenes
3D Generative Models
Quantum Algorithms for Computer Vision and Graphics
Discovering Fatigued Movements for Virtual Character Animation.
N. Cheema, R, Xu, N. H. Kim, P. Hämäläinen, V. Golyanik, M. Habermann, C. Theobalt, P. Slusallek.
SIGGRAPH ASIA, 2023 (conference paper).
[arxiv]
[project page]
Description:
AvatarStudio: Text-driven Editing of 3D Dynamic Human Head Avatars.
M. Mendiratta, X. Pan, M. Elgharib, K. Teotia, Mallikarjun B R, A. Tewari, V. Golyanik, A. Kortylewski and C. Theobalt.
SIGGRAPH ASIA, 2023.
[arxiv]
[project page]
Description:
We propose a text-based method for editing the appearance of a dynamic full head avatar. Our approach edits neural radiance field with a text-to-image diffusion model. Our optimization strategy incorporates multiple keyframes representing different camera viewpoints and time stamps of a video performance into a single diffusion model.
EgoLocate: Real-time Motion Capture, Localization, and Mapping with Sparse Body-mounted Sensors.
X. Yi, Y. Zhou, M. Habermann, V. Golyanik, S. Pan, C. Theobalt and F. Xu.
SIGGRAPH, 2023.
[arxiv]
[project page]
[source code]
Description:
EgoLocate is a system that simultaneously performs human motion capture (mocap), localization, and mapping in real time from sparse body-mounted sensors, including six inertial measurement units (IMUs) and a monocular phone camera.
Quantum Multi-Model Fitting.
M. Farina, L. Magri, W. Menapace, E. Ricci, V. Golyanik and F. Arrigoni.
Computer Vision and Pattern Recognition (CVPR), 2023; CVPR Highlight (selected 10%).
[paper]
[project page]
[source code]
Description:
We formulate multi-model fitting (MMF) as a problem that can be efficiently sampled by modern adiabatic quantum computers without the relaxation of the objective function. We also propose an iterative and decomposed version of our method, which supports real-world-sized problems. The experimental evaluation demonstrates promising results on a variety of datasets.
MoFusion: A Framework for Denoising-Diffusion-based Motion Synthesis.
R. Dabral, M. H. Mughal, V. Golyanik and C. Theobalt.
Computer Vision and Pattern Recognition (CVPR), 2023; CVPR Highlight (selected 10%).
[paper]
[project page]
[bibtex]
Description:
We introduce MoFusion, i.e., a new denoising-diffusion-based framework for high-quality conditional human motion synthesis that can synthesise long, temporally plausible, and semantically accurate motions based on a range of conditioning contexts (such as music and text). We also present ways to introduce well-known kinematic losses for motion plausibility within the motion diffusion framework through our scheduled weighting strategy. The learned latent space can be used for several interactive motion-editing applications like in-betweening, seed-conditioning, and text-based editing, thus, providing crucial abilities for virtual-character animation and robotics.
EventNeRF: Neural Radiance Fields from a Single Colour Event Camera.
V. Rudnev, M. Elgharib, C. Theobalt, V. Golyanik.
Computer Vision and Pattern Recognition (CVPR), 2023.
[paper]
[project page]
[bibtex]
Description:
This pre-print proposes the first approach for 3D-consistent, dense and photorealistic novel view synthesis using just a single colour event stream as input. At the core of our method is a neural radiance field trained entirely in a self-supervised manner from events while preserving the original resolution of the colour event channels. Next, our ray sampling strategy is tailored to events and allows for data-efficient training. At test, our method produces results in the RGB space at unprecedented quality.
CCuantuMM: Cycle-Consistent Quantum-Hybrid Matching of Multiple Shapes.
H. Bhatia, E. Tretschk, Z. Lähner, M. Seelbach Benkner, M. Moeller, C. Theobalt and V. Golyanik.
Computer Vision and Pattern Recognition (CVPR), 2023.
[paper]
[project page]
[source code]
Description:
Jointly matching multiple, non-rigidly deformed 3D shapes is a challenging, NP-hard problem. A perfect matching is necessarily cycle-consistent: Following the pairwise point correspondences along several shapes must end up at the starting vertex of the original shape. This paper introduces the first quantum-hybrid approach for 3D shape multi-matching. Its iterative formulation is admissible to modern adiabatic quantum hardware and scales linearly with the total number of input shapes. On benchmark datasets, the proposed approach significantly outperforms extensions to multi-shape matching of a previous quantum-hybrid two-shape matching method and is on-par with classical multi-matching methods.
Self-supervised Pre-training with Masked Shape Prediction for 3D Scene Understanding.
L. Jiang, Z. Yang, S. Shi, V. Golyanik, D. Dai and B. Schiele.
Computer Vision and Pattern Recognition (CVPR), 2023.
[paper]
Description:
This paper introduces Masked Shape Prediction (MSP), a new framework to conduct masked signal modeling in 3D scenes. MSP uses the essential 3D semantic cue, i.e., geometric shape, as the prediction target for masked points. The context-enhanced shape target consisting of explicit shape context and implicit deep shape feature is proposed to facilitate exploiting contextual cues in shape prediction. Meanwhile, the pre-training architecture in MSP is carefully designed to alleviate the masked shape leakage from point coordinates. Experiments on multiple 3D understanding tasks on both indoor and outdoor datasets demonstrate the effectiveness of MSP in learning good feature representations to consistently boost downstream performance.
Unbiased 4D: Monocular 4D Reconstruction with a Neural Deformation Model.
E. C. M. Johnson, M. Habermann, S. Shimada, V. Golyanik and C. Theobalt.
Computer Vision and Pattern Recognition (CVPR) Workshops, 2023.
[paper]
[project page]
[source code]
[data]
[bibtex]
Description:
Our method, Ub4D, handles large deformations, performs shape completion in occluded regions, and can operate on monocular RGB videos directly by using differentiable volume rendering. This technique includes three new---in the context of non-rigid 3D reconstruction---components, i.e., 1) A coordinate-based and implicit neural representation for non-rigid scenes, which in conjunction with differentiable volume rendering enables an unbiased reconstruction of dynamic scenes, 2) a proof that extends the unbiased formulation of volume rendering to dynamic scenes, and 3) a novel dynamic scene flow loss, which enables the reconstruction of larger deformations by leveraging the coarse estimates of other methods.
QuAnt: Quantum Annealing with Learnt Couplings.
M. Seelbach Benkner, M. Krahn, E. Tretschk, Z. Lähner, M. Moeller and V. Golyanik.
International Conference on Learning Representations (ICLR), 2023; Oral (top 25% of the accepted papers).
[project page]
[paper]
[bibtex]
Abstract:
This paper proposes to learn QUBO forms from data through gradient backpropagation instead of deriving them. As a result, the solution encodings can be chosen flexibly and compactly. Furthermore, our methodology is general and virtually independent of the specifics of the target problem type. We demonstrate the advantages of learnt QUBOs on the diverse problem types of graph matching, 2D point cloud alignment and 3D rotation estimation.
State of the Art in Dense Monocular Non-Rigid 3D Reconstruction.
E. Tretschk*, N. Kairanda*, M. B R, R. Dabral, A. Kortylewski, B. Egger, M. Habermann, P. Fua, C. Theobalt and V. Golyanik.
* equal contribution.
Eurographics 2023 (Full STARs).
[draft]
[project page]
[bibtex]
Description:
This survey focuses on state-of-the-art methods for dense non-rigid 3D reconstruction of various deformable objects and composite scenes from monocular videos or sets of monocular views. It reviews the fundamentals of 3D reconstruction and deformation modeling from 2D image observations. We then start from general methods—that handle arbitrary scenes and make only a few prior assumptions—and proceed towards techniques making stronger assumptions about the observed objects and types of deformations (e.g. human faces, bodies, hands, and animals). A significant part of this STAR is also devoted to classification and a high-level comparison of the methods, as well as an overview of the datasets for training and evaluation of the discussed techniques. We conclude by discussing open challenges in the field and the social aspects associated with the usage of the reviewed methods.
Scene-Aware 3D Multi-Human Motion Capture from a Single Camera.
D. Luvizon, M. Habermann, V. Golyanik, A. Kortylewski and C. Theobalt.
Eurographics 2023.
[paper]
[project page]
Description:
We introduce the first non-linear optimization-based approach that jointly solves for the absolute 3D position of each human, their articulated pose, their individual shapes as well as the scale of the scene. Given the per-frame 3D estimates of the humans and scene point-cloud, we perform a space-time coherent optimization over the video to ensure temporal, spatial and physical plausibility. We consistently outperform previous methods and we qualitatively demonstrate that our method is robust to in-the-wild conditions including challenging scenes with people of different sizes.
IMoS: Intent-Driven Full-Body Motion Synthesis for Human-Object Interactions.
A. Ghosh, R. Dabral, V. Golyanik, C. Theobalt and P. Slusallek.
Eurographics 2023.
[project page]
[paper]
[bibtex]
Description:
We synthesize the full-body pose sequences along with the 3D object positions from textual inputs. Our method can synthesize single-handed as well as two-handed interactions depending on the intent and the type of the object used.
2022
HiFECap: Monocular High-Fidelity and Expressive Capture of Human Performances.
Y. Jiang, M. Habermann, V. Golyanik and C. Theobalt.
British Machine Vision Conference (BMVC), 2022.
[project page]
[paper]
[bibtex]
Description:
We propose HiFECap, a new neural human performance capture approach, which simultaneously captures human pose, clothing, facial expression, and hands just from a single RGB video. We demonstrate that our proposed network architecture, the carefully designed training strategy, and the tight integration of parametric face and hand models to a template mesh enable the capture of all these individual aspects. Importantly, our method also captures high-frequency details, such as deforming wrinkles on the clothes, better than the previous works.
Generation of Truly Random Numbers on a Quantum Annealer.
H. Bhatia, E. Tretschk, C. Theobalt and V. Golyanik.
More details coming soon.
[project page]
[paper]
[bibtex]
Description:
We discuss the observed qubits' properties and their influence on the random number generation and consider various physical factors that influence the performance of our generator, i.e., digital-to-analogue quantisation errors, flux errors, temperature errors and spin bath polarisation. The numbers generated by the proposed algorithm successfully pass various tests on randomness from the NIST test suite.
Q-FW: A Hybrid Classical-Quantum Frank-Wolfe for Quadratic Binary Optimization.
A. Yurtsever, T. Birdal and V. Golyanik.
European Conference on Computer Vision (ECCV), 2022.
[project page]
[paper]
[bibtex]
[poster]
Description:
We present a hybrid classical-quantum framework based on the Frank-Wolfe algorithm, Q-FW, for solving quadratic, linearly-constrained, binary optimization problems on quantum annealers (QA). Q-FW first reformulates constrained-QBO as a copositive program (CP), then employs Frank-Wolfe iterations to solve CP while satisfying linear (in)equality constraints. This procedure unrolls the original constrained-QBO into a set of unconstrained QUBOs all of which are solved, in a sequel, on a QA. We use D-Wave Advantage QA to conduct synthetic and real experiments on two important computer vision problems, graph matching and permutation synchronization, which demonstrate that our approach is effective in alleviating the need for an explicit regularization coefficient.
UnrealEgo: A New Dataset for Robust Egocentric 3D Human Motion Capture.
H. Akada, J. Wang, S. Shimada, M. Takahashi, C. Theobalt and V. Golyanik.
European Conference on Computer Vision (ECCV), 2022.
[project page]
[paper]
[bibtex]
Description:
UnrealEgo is a large-scale naturalistic dataset for egocentric 3D human pose estimation. It is based on an advanced concept of eyeglasses equipped with two fisheye cameras that can be used in unconstrained environments. The experiments show that our simple yet effective approach for egocentric 3D human motion capture outperforms the previous methods.
Quantum Motion Segmentation.
F. Arrigoni, W. Menapace, M. Seelbach Benkner, E. Ricci and V. Golyanik.
European Conference on Computer Vision (ECCV), 2022.
[project page]
[paper]
[bibtex]
Abstract:
Motion segmentation is a challenging problem that seeks to identify independent motions in two or several input images. This paper introduces the first algorithm for motion segmentation that relies on adiabatic quantum optimization of the objective function. The proposed method achieves on-par performance with the state of the art on problem instances which can be mapped to modern quantum annealers.
HULC: 3D HUman Motion Capture with Pose Manifold Sampling and Dense Contact Guidance.
S. Shimada, V. Golyanik, Z. Li, P. Pérez, W. Xu and C. Theobalt.
European Conference on Computer Vision (ECCV), 2022.
[paper]
[project page]
Neural Radiance Fields for Outdoor Scene Relighting.
V. Rudnev, M. Elgharib, W. Smith, L. Liu, V. Golyanik and C. Theobalt.
European Conference on Computer Vision (ECCV), 2022.
[paper]
[project page]
[bibtex]
MoCapDeform: Monocular 3D Human Motion Capture in Deformable Scenes.
Z. Li, S. Shimada, B. Schiele, C. Theobalt and V. Golyanik.
International Conference on 3D Vision (3DV), 2022; Oral.
Best Student Paper Award.
[project page]
[paper]
[bibtex]
Description:
Our MoCapDeform algorithm is the first that models non-rigid scene deformations and finds the accurate global 3D poses of the subject by human-deformable scene interaction constraints, achieving increased accuracy with significantly fewer penetrations.
φ-SfT: Shape-from-Template with a Physics-Based Deformation Model.
N. Kairanda, E. Tretschk, M. Elgharib, C. Theobalt and V. Golyanik.
Computer Vision and Pattern Recognition (CVPR), 2022.
[paper]
[project page]
[source code]
[bibtex]
Playable Environments: Video Manipulation in Space and Time.
W. Menapace, S. Lathuilière*, A. Siarohin, C. Theobalt*, S. Tulyakov*, V. Golyanik*, and E. Ricci*.
* equal senior contribution.
Computer Vision and Pattern Recognition (CVPR), 2022.
[paper]
[project page]
[github]
[bibtex]
Advances in Neural Rendering.
A. Tewari*, J. Thies*, B. Mildenhall*, P. Srinivasan*, E. Tretschk, Y. Wang, C. Lassner, V. Sitzmann, R. Martin-Brualla, S. Lombardi, C. Theobalt, M. Niessner, J. T. Barron, G. Wetzstein, M. Zollhöfer and V. Golyanik.
* equal contribution.
State of the Art Report at Eurographics 2022.
[paper]
[project page]
[bibtex]
2021
Convex Joint Graph Matching and Clustering via Semidefinite Relaxations.
M. Krahn, F. Bernard and V. Golyanik.
International Conference on 3D Vision (3DV), 2021.
[paper]
[project page]
[bibtex]
HumanGAN: A Generative Model of Human Images.
K. Sarkar, L. Liu, V. Golyanik, and C. Theobalt.
International Conference on 3D Vision (3DV), 2021; Oral
[paper]
[project page]
[bibtex]
HandVoxNet++: 3D Hand Shape and Pose Estimation using Voxel-Based Neural Networks.
J. Malik, S. Shimada, A. Elhayek, S. A. Ali, C. Theobalt, V. Golyanik and D. Stricker.
Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2021.
[IEEE Xplore]
[arXiv.org]
[project page]
[bibtex]
Gravity-Aware 3D Human-Object Reconstruction.
R. Dabral, S. Shimada, A. Jain, C. Theobalt and V. Golyanik.
International Conference on Computer Vision (ICCV), 2021.
[paper]
[project page]
[bibtex]
Q-Match: Iterative Shape Matching via Quantum Annealing.
M. Seelbach Benkner, Z. Lähner, V. Golyanik, C. Wunderlich, C. Theobalt and M. Moeller.
International Conference on Computer Vision (ICCV), 2021.
[paper]
[project page]
[bibtex]
EventHands: Real-Time Neural 3D Hand Pose Estimation from an Event Stream.
V. Rudnev, V. Golyanik, J. Wang, H.-P. Seidel, F. Mueller, M. Elgharib and C. Theobalt.
International Conference on Computer Vision (ICCV), 2021.
Featured in "ICCV 2021 Daily" (link)
[paper]
[project page]
[source code]
[bibtex]
Non-Rigid Neural Radiance Fields: Reconstruction and Novel View Synthesis of a Deforming Scene from Monocular Video.
E. Tretschk, A. Tewari, V. Golyanik, M. Zollhöfer, C. Lassner and C. Theobalt.
International Conference on Computer Vision (ICCV), 2021.
[paper]
[project page]
[source code]
[bibtex]
Neural Monocular 3D Human Motion Capture with Physical Awareness.("Neural PhysCap")
S. Shimada, V. Golyanik, W. Xu, P. Pérez and C. Theobalt.
SIGGRAPH, 2021.
[paper]
[arXiv]
[bibtex]
[project page]
[source code]
Quantum Permutation Synchronization.T. Birdal*, V. Golyanik*, C. Theobalt and L. Guibas.
* equal contribution.
Computer Vision and Pattern Recognition (CVPR), 2021.
[paper]
[arXiv]
[project page]
[project page (github)]
[bibtex]
High-Fidelity Neural Human Motion Transfer from Monocular Video.
M. Kappel, V. Golyanik, M. Elgharib, J.-O. Henningson, H.-P. Seidel, S. Castillo, C. Theobalt and M. Magnor.
Computer Vision and Pattern Recognition (CVPR), 2021; Oral.
[paper]
[project page]
[bibtex]
[source code]
Pose-Guided Human Animation from a Single Image in the Wild.
J. S. Yoon, L. Liu, V. Golyanik, K. Sarkar, H. S. Park, and C. Theobalt.
Computer Vision and Pattern Recognition (CVPR), 2021.
[paper]
[project page]
[video]
[bibtex]
Differentiable Event Stream Simulator for Non-Rigid 3D Tracking.
J. Nehvi, V. Golyanik, F. Mueller, H.-P. Seidel, M. Elgharib and C. Theobalt.
CVPR Workshop on Event-based Vision, 2021.
[paper]
[arXiv]
[supplement]
[source code (upon request)]
[bibtex]
[project page]
Fast Gravitational Approach for Rigid Point Set Registration with Ordinary Differential Equations.
S. A. Ali, K. Kahraman, C. Theobalt, D. Stricker and V. Golyanik.
IEEE Access, 2021.
[paper]
[arXiv]
[project page]
[bibtex]
2020
PhysCap: Physically Plausible Monocular 3D Motion Capture in Real Time.
S. Shimada, V. Golyanik, W. Xu and C. Theobalt.
SIGGRAPH Asia, 2020.
[paper (arXiv.org)]
[bibtex]
[project page]
Egocentric Videoconferencing.
M. Mendiratta*, M. Elgharib*, J. Thies, M. Nießner, H.-P. Seidel, A. Tewari, V. Golyanik and C. Theobalt.
* equal contribution.
SIGGRAPH Asia, 2020.
[draft]
[supplement]
[bibtex]
[project page]
Fast Simultaneous Gravitational Alignment of Multiple Point Sets.V. Golyanik, S. Shimada and C. Theobalt.
3DV, 2020; Oral.
[draft]
[bibtex]
[project page]
Adiabatic Quantum Graph Matching with Permutation Matrix Constraints.
M. Seelbach Benkner, V. Golyanik, C. Theobalt and M. Moeller.
3DV, 2020.
[draft]
[supplement]
[bibtex]
[project page]
Neural Dense Non-Rigid Structure from Motion with Latent Space Constraints.
V. Sidhu, E. Tretschk, V. Golyanik, A. Agudo and C. Theobalt.
European Conference on Computer Vision (ECCV), 2020.
[paper]
[supplement]
[video]
[bibtex]
[project page (includes the ECCV talks)]
HTML: A Parametric Hand Texture Model for 3D Hand Reconstruction and Personalization.
N. Qian, J. Wang, F. Müller, F. Bernard, V. Golyanik and C. Theobalt.
European Conference on Computer Vision (ECCV), 2020.
[paper]
[supplement]
[video]
[bibtex]
[project page]
DEMEA: Deep Mesh Autoencoders for Non-Rigidly Deforming Objects.
E. Tretschk, A. Tewari, M. Zollhöfer, V. Golyanik and C. Theobalt.
European Conference on Computer Vision (ECCV), 2020; Spotlight
[paper]
[supplement]
[bibtex]
[arXiv]
[project page (includes the ECCV talks)]
A Quantum Computational Approach to Correspondence Problems on Point Sets.V. Golyanik and C. Theobalt.
In Computer Vision and Pattern Recognition (CVPR), 2020.
[paper]
[slides]
[poster]
[bibtex]
[arXiv]
[project page]
EventCap: Monocular 3D Capture of High-Speed Human Motions using an Event Camera.
L. Xu, W. Xu, V. Golyanik, M. Habermann, L. Fang and C. Theobalt.
In Computer Vision and Pattern Recognition (CVPR), 2020; Oral
[paper]
[supplement]
[bibtex]
[arXiv]
[project page]
HandVoxNet: Deep Voxel-Based Network for 3D Hand Shape and Pose Estimation from a Single Depth Map.
J. Malik, I. Abdelaziz, A. Elhayek, S. Shimada, S. A. Ali, V. Golyanik, C. Theobalt and D. Stricker.
In Computer Vision and Pattern Recognition (CVPR), 2020.
[paper]
[supplement]
[bibtex]
[arXiv]
[project page]
Structure from Articulated Motion: Accurate and Stable Monocular 3D Reconstruction without Training Data.
O. Kovalenko, V. Golyanik, J. Malik, A. Elhayek and D. Stricker.
Sensors (Volume 19, Issue 20), 2019.
[paper]
[project page]
Accelerated Gravitational Point Set Alignment with Altered Physical Laws.V. Golyanik, C. Theobalt and D. Stricker.
International Conference on Computer Vision (ICCV), 2019;
[paper]
[supplementary material]
[poster]
[bibtex]
[project page]
DispVoxNets: Non-Rigid Point Set Alignment with Supervised Learning Proxies.
S. Shimada, V. Golyanik, E. Tretschk, D. Stricker and C. Theobalt.
In International Conference on 3D Vision (3DV), 2019; Oral
[paper]
[poster]
[presentation]
[project page]
[arXiv]
[bibtex]
IsMo-GAN: Adversarial Learning for Monocular Non-Rigid 3D Reconstruction.
S. Shimada, V. Golyanik, C. Theobalt and D. Stricker.
Computer Vision and Pattern Recognition Workshops (Photogrammetric Computer Vision Workshop), 2019; Oral
[paper]
[bibtex]
[arXiv]
[project page]
Consolidating Segmentwise Non-Rigid Structure from Motion. V. Golyanik, A. Jonas and D. Stricker.
Machine Vision Applications (MVA), 2019; Oral
[paper]
[project page]
[bibtex]
NRGA: Gravitational Approach for Non-Rigid Point Set Registration.
S. A. Ali. V. Golyanik and D. Stricker.
International Conference on 3D Vision (3DV), 2018; Oral
[paper]
[Supplementary Video (Download, YouTube)]
[poster]
[bibtex]
HDM-Net: Monocular Non-Rigid 3D Reconstruction with Learned Deformation Model. V. Golyanik, S. Shimada, K. Varanasi and D. Stricker.
EuroVR, 2018; Oral (Long Paper)
[paper]
[HDM-Net data set]
[bibtex]
Multiframe Scene Flow with Piecewise Rigid Motion. V. Golyanik, K. Kim, R. Maier, M. Nießner, D. Stricker and J. Kautz.
International Conference on 3D Vision (3DV), 2017; Spotlight Oral
[paper]
[arXiv]
[supplementary material]
[poster]
[bibtex]
Scalable Dense Monocular Surface Reconstruction.
M.D.Ansari, V. Golyanik and D. Stricker.
International Conference on 3D Vision (3DV), 2017.
[paper]
[arXiv]
[bibtex]
A Framework for an Accurate Point Cloud Based Registration of Full 3D Human Body Scans. V. Golyanik, G. Reis, B. Taetz and D. Stricker.
Machine Vision Applications (MVA), 2017.
[paper]
[bibtex]
Dense Batch Non-Rigid Structure from Motion in a Second. V. Golyanik and D. Stricker.
Winter Conference on Applications of Computer Vision (WACV), 2017.
[paper]
[supplementary video]
[poster]
[bibtex]
Accurate 3D Reconstruction of Dynamic Scenes from Monocular Image Sequences with Severe Occlusions. V. Golyanik, T. Fetzer and D. Stricker.
Winter Conference on Applications of Computer Vision (WACV), 2017.
[paper]
[supplementary material]
[poster]
[arXiv]
[bibtex]
Gravitational Approach for Point Set Registration. V. Golyanik, S. A. Ali and D. Stricker.
Computer Vision and Pattern Recognition (CVPR), 2016.
[paper]
[supplementary material]
[bibtex]
Extended Coherent Point Drift Algorithm with Correspondence Priors and Optimal Subsampling. V. Golyanik, B. Taetz, G. Reis and D. Stricker.
Winter Conference on Applications of Computer Vision (WACV), 2016.
[paper]
[poster]
[bibtex]
[WACV Talk]
Occlusion-Aware Video Registration for Highly Non-Rigid Objects.
B. Taetz, G. Bleser, V. Golyanik and D. Stricker.
Winter Conference on Applications of Computer Vision (WACV), 2016.
Best Paper Award.
[paper]
[supplementary material]
[bibtex]
[WACV Talk]