47 results on '"Ploumpis, Stylianos"'
Search Results
2. ShapeFusion: A 3D diffusion model for localized shape editing
- Author
-
Potamias, Rolandos Alexandros, Tarasiou, Michail, Ploumpis, Stylianos, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
In the realm of 3D computer vision, parametric models have emerged as a ground-breaking methodology for the creation of realistic and expressive 3D avatars. Traditionally, they rely on Principal Component Analysis (PCA), given its ability to decompose data to an orthonormal space that maximally captures shape variations. However, due to the orthogonality constraints and the global nature of PCA's decomposition, these models struggle to perform localized and disentangled editing of 3D shapes, which severely affects their use in applications requiring fine control such as face sculpting. In this paper, we leverage diffusion models to enable diverse and fully localized edits on 3D meshes, while completely preserving the un-edited regions. We propose an effective diffusion masking training strategy that, by design, facilitates localized manipulation of any shape region, without being limited to predefined regions or to sparse sets of predefined control vertices. Following our framework, a user can explicitly set their manipulation region of choice and define an arbitrary set of vertices as handles to edit a 3D mesh. Compared to the current state-of-the-art our method leads to more interpretable shape manipulations than methods relying on latent code state, greater localization and generation diversity while offering faster inference than optimization based approaches. Project page: https://rolpotamias.github.io/Shapefusion/, Comment: Project Page: https://rolpotamias.github.io/Shapefusion/
- Published
- 2024
3. AnimateMe: 4D Facial Expressions via Diffusion Models
- Author
-
Gerogiannis, Dimitrios, Papantoniou, Foivos Paraperas, Potamias, Rolandos Alexandros, Lattas, Alexandros, Moschoglou, Stylianos, Ploumpis, Stylianos, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
The field of photorealistic 3D avatar reconstruction and generation has garnered significant attention in recent years; however, animating such avatars remains challenging. Recent advances in diffusion models have notably enhanced the capabilities of generative models in 2D animation. In this work, we directly utilize these models within the 3D domain to achieve controllable and high-fidelity 4D facial animation. By integrating the strengths of diffusion processes and geometric deep learning, we employ Graph Neural Networks (GNNs) as denoising diffusion models in a novel approach, formulating the diffusion process directly on the mesh space and enabling the generation of 3D facial expressions. This facilitates the generation of facial deformations through a mesh-diffusion-based model. Additionally, to ensure temporal coherence in our animations, we propose a consistent noise sampling method. Under a series of both quantitative and qualitative experiments, we showcase that the proposed method outperforms prior work in 4D expression synthesis by generating high-fidelity extreme expressions. Furthermore, we applied our method to textured 4D facial expression generation, implementing a straightforward extension that involves training on a large-scale textured 4D facial expression database.
- Published
- 2024
4. Locally Adaptive Neural 3D Morphable Models
- Author
-
Tarasiou, Michail, Potamias, Rolandos Alexandros, O'Sullivan, Eimear, Ploumpis, Stylianos, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
We present the Locally Adaptive Morphable Model (LAMM), a highly flexible Auto-Encoder (AE) framework for learning to generate and manipulate 3D meshes. We train our architecture following a simple self-supervised training scheme in which input displacements over a set of sparse control vertices are used to overwrite the encoded geometry in order to transform one training sample into another. During inference, our model produces a dense output that adheres locally to the specified sparse geometry while maintaining the overall appearance of the encoded object. This approach results in state-of-the-art performance in both disentangling manipulated geometry and 3D mesh reconstruction. To the best of our knowledge LAMM is the first end-to-end framework that enables direct local control of 3D vertex geometry in a single forward pass. A very efficient computational graph allows our network to train with only a fraction of the memory required by previous methods and run faster during inference, generating 12k vertex meshes at $>$60fps on a single CPU thread. We further leverage local geometry control as a primitive for higher level editing operations and present a set of derivative capabilities such as swapping and sampling object parts. Code and pretrained models can be found at https://github.com/michaeltrs/LAMM., Comment: 10 pages, 9 figures, 2 tables
- Published
- 2024
5. FitMe: Deep Photorealistic 3D Morphable Model Avatars
- Author
-
Lattas, Alexandros, Moschoglou, Stylianos, Ploumpis, Stylianos, Gecer, Baris, Deng, Jiankang, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Graphics ,Computer Science - Machine Learning ,I.2.10 ,I.3.7 ,I.4.1 - Abstract
In this paper, we introduce FitMe, a facial reflectance model and a differentiable rendering optimization pipeline, that can be used to acquire high-fidelity renderable human avatars from single or multiple images. The model consists of a multi-modal style-based generator, that captures facial appearance in terms of diffuse and specular reflectance, and a PCA-based shape model. We employ a fast differentiable rendering process that can be used in an optimization pipeline, while also achieving photorealistic facial shading. Our optimization process accurately captures both the facial reflectance and shape in high-detail, by exploiting the expressivity of the style-based latent representation and of our shape model. FitMe achieves state-of-the-art reflectance acquisition and identity preservation on single "in-the-wild" facial images, while it produces impressive scan-like results, when given multiple unconstrained facial images pertaining to the same identity. In contrast with recent implicit avatar reconstructions, FitMe requires only one minute and produces relightable mesh and texture-based avatars, that can be used by end-user applications., Comment: Accepted at CVPR 2023, project page at https://lattas.github.io/fitme , 17 pages including supplementary material
- Published
- 2023
6. Dynamic Neural Portraits
- Author
-
Doukas, Michail Christos, Ploumpis, Stylianos, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
We present Dynamic Neural Portraits, a novel approach to the problem of full-head reenactment. Our method generates photo-realistic video portraits by explicitly controlling head pose, facial expressions and eye gaze. Our proposed architecture is different from existing methods that rely on GAN-based image-to-image translation networks for transforming renderings of 3D faces into photo-realistic images. Instead, we build our system upon a 2D coordinate-based MLP with controllable dynamics. Our intuition to adopt a 2D-based representation, as opposed to recent 3D NeRF-like systems, stems from the fact that video portraits are captured by monocular stationary cameras, therefore, only a single viewpoint of the scene is available. Primarily, we condition our generative model on expression blendshapes, nonetheless, we show that our system can be successfully driven by audio features as well. Our experiments demonstrate that the proposed method is 270 times faster than recent NeRF-based reenactment methods, with our networks achieving speeds of 24 fps for resolutions up to 1024 x 1024, while outperforming prior works in terms of visual quality., Comment: In IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) 2023
- Published
- 2022
7. AvatarMe++: Facial Shape and BRDF Inference with Photorealistic Rendering-Aware GANs
- Author
-
Lattas, Alexandros, Moschoglou, Stylianos, Ploumpis, Stylianos, Gecer, Baris, Ghosh, Abhijeet, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Graphics ,I.4.1 ,I.3.7 ,I.2.10 - Abstract
Over the last years, many face analysis tasks have accomplished astounding performance, with applications including face generation and 3D face reconstruction from a single "in-the-wild" image. Nevertheless, to the best of our knowledge, there is no method which can produce render-ready high-resolution 3D faces from "in-the-wild" images and this can be attributed to the: (a) scarcity of available data for training, and (b) lack of robust methodologies that can successfully be applied on very high-resolution data. In this work, we introduce the first method that is able to reconstruct photorealistic render-ready 3D facial geometry and BRDF from a single "in-the-wild" image. We capture a large dataset of facial shape and reflectance, which we have made public. We define a fast facial photorealistic differentiable rendering methodology with accurate facial skin diffuse and specular reflection, self-occlusion and subsurface scattering approximation. With this, we train a network that disentangles the facial diffuse and specular BRDF components from a shape and texture with baked illumination, reconstructed with a state-of-the-art 3DMM fitting method. Our method outperforms the existing arts by a significant margin and reconstructs high-resolution 3D faces from a single low-resolution image, that can be rendered in various applications, and bridge the uncanny valley., Comment: Project and Dataset page: ( https://github.com/lattas/AvatarMe ). 20 pages, including supplemental materials. Accepted for publishing at IEEE Transactions on Pattern Analysis and Machine Intelligence on 13 November 2021. Copyright 2021 IEEE. Personal use of this material is permitted
- Published
- 2021
- Full Text
- View/download PDF
8. 3D human tongue reconstruction from single 'in-the-wild' images
- Author
-
Ploumpis, Stylianos, Moschoglou, Stylianos, Triantafyllou, Vasileios, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence ,Computer Science - Graphics - Abstract
3D face reconstruction from a single image is a task that has garnered increased interest in the Computer Vision community, especially due to its broad use in a number of applications such as realistic 3D avatar creation, pose invariant face recognition and face hallucination. Since the introduction of the 3D Morphable Model in the late 90's, we witnessed an explosion of research aiming at particularly tackling this task. Nevertheless, despite the increasing level of detail in the 3D face reconstructions from single images mainly attributed to deep learning advances, finer and highly deformable components of the face such as the tongue are still absent from all 3D face models in the literature, although being very important for the realness of the 3D avatar representations. In this work we present the first, to the best of our knowledge, end-to-end trainable pipeline that accurately reconstructs the 3D face together with the tongue. Moreover, we make this pipeline robust in "in-the-wild" images by introducing a novel GAN method tailored for 3D tongue surface generation. Finally, we make publicly available to the community the first diverse tongue dataset, consisting of 1,800 raw scans of 700 individuals varying in gender, age, and ethnicity backgrounds. As we demonstrate in an extensive series of quantitative as well as qualitative experiments, our model proves to be robust and realistically captures the 3D tongue structure, even in adverse "in-the-wild" conditions., Comment: 10 pages, 9 figures
- Published
- 2021
9. Fast-GANFIT: Generative Adversarial Network for High Fidelity 3D Face Reconstruction
- Author
-
Gecer, Baris, Ploumpis, Stylianos, Kotsia, Irene, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
A lot of work has been done towards reconstructing the 3D facial structure from single images by capitalizing on the power of Deep Convolutional Neural Networks (DCNNs). In the recent works, the texture features either correspond to components of a linear texture space or are learned by auto-encoders directly from in-the-wild images. In all cases, the quality of the facial texture reconstruction is still not capable of modeling facial texture with high-frequency details. In this paper, we take a radically different approach and harness the power of Generative Adversarial Networks (GANs) and DCNNs in order to reconstruct the facial texture and shape from single images. That is, we utilize GANs to train a very powerful facial texture prior \edit{from a large-scale 3D texture dataset}. Then, we revisit the original 3D Morphable Models (3DMMs) fitting making use of non-linear optimization to find the optimal latent parameters that best reconstruct the test image but under a new perspective. In order to be robust towards initialisation and expedite the fitting process, we propose a novel self-supervised regression based approach. We demonstrate excellent results in photorealistic and identity preserving 3D face reconstructions and achieve for the first time, to the best of our knowledge, facial texture reconstruction with high-frequency details., Comment: TPAMI camera ready (submitted: 05-May-2020); Check project page: https://github.com/barisgecer/GANFit. arXiv admin note: substantial text overlap with arXiv:1902.05978
- Published
- 2021
- Full Text
- View/download PDF
10. Learning to Generate Customized Dynamic 3D Facial Expressions
- Author
-
Potamias, Rolandos Alexandros, Zheng, Jiali, Ploumpis, Stylianos, Bouritsas, Giorgos, Ververas, Evangelos, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Recent advances in deep learning have significantly pushed the state-of-the-art in photorealistic video animation given a single image. In this paper, we extrapolate those advances to the 3D domain, by studying 3D image-to-video translation with a particular focus on 4D facial expressions. Although 3D facial generative models have been widely explored during the past years, 4D animation remains relatively unexplored. To this end, in this study we employ a deep mesh encoder-decoder like architecture to synthesize realistic high resolution facial expressions by using a single neutral frame along with an expression identification. In addition, processing 3D meshes remains a non-trivial task compared to data that live on grid-like structures, such as images. Given the recent progress in mesh processing with graph convolutions, we make use of a recently introduced learnable operator which acts directly on the mesh structure by taking advantage of local vertex orderings. In order to generalize to 4D facial expressions across subjects, we trained our model using a high resolution dataset with 4D scans of six facial expressions from 180 subjects. Experimental results demonstrate that our approach preserves the subject's identity information even for unseen subjects and generates high quality expressions. To the best of our knowledge, this is the first study tackling the problem of 4D facial expression synthesis., Comment: accepted at European Conference on Computer Vision 2020 (ECCV)
- Published
- 2020
11. AvatarMe: Realistically Renderable 3D Facial Reconstruction 'in-the-wild'
- Author
-
Lattas, Alexandros, Moschoglou, Stylianos, Gecer, Baris, Ploumpis, Stylianos, Triantafyllou, Vasileios, Ghosh, Abhijeet, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Graphics ,I.2.10 ,I.3.7 ,I.4.1 - Abstract
Over the last years, with the advent of Generative Adversarial Networks (GANs), many face analysis tasks have accomplished astounding performance, with applications including, but not limited to, face generation and 3D face reconstruction from a single "in-the-wild" image. Nevertheless, to the best of our knowledge, there is no method which can produce high-resolution photorealistic 3D faces from "in-the-wild" images and this can be attributed to the: (a) scarcity of available data for training, and (b) lack of robust methodologies that can successfully be applied on very high-resolution data. In this paper, we introduce AvatarMe, the first method that is able to reconstruct photorealistic 3D faces from a single "in-the-wild" image with an increasing level of detail. To achieve this, we capture a large dataset of facial shape and reflectance and build on a state-of-the-art 3D texture and shape reconstruction method and successively refine its results, while generating the per-pixel diffuse and specular components that are required for realistic rendering. As we demonstrate in a series of qualitative and quantitative experiments, AvatarMe outperforms the existing arts by a significant margin and reconstructs authentic, 4K by 6K-resolution 3D faces from a single low-resolution image that, for the first time, bridges the uncanny valley., Comment: Accepted to CVPR2020. Project page: github.com/lattas/AvatarMe with high resolution results, data and more. 10 pages, 9 figures
- Published
- 2020
12. 3D head morphable models and beyond : algorithms and applications
- Author
-
Ploumpis, Stylianos A. and Zafeiriou, Stefanos
- Abstract
It has been more than 20 year since the introduction of 3D morphable models (3DMM) in the computer vision literature. They were proposed as a face representation based on principal components analysis for the task of image analysis, photorealist-manipulation, and 3D reconstruction from single images. Even so, to this date, the applications of such models are limited by a number of factors. Firstly, training correctly 3DMMs require a vast amount of 3D data that most of the times are not publicly available to the research community due to increasingly stringent data protection regulations. Hence, it is extremely difficult to combine and enrich multiple attributes of the human face/head without the initial 3D images. Additionally, many 3DMMs utilize different templates that describe distinct parts of the human face/head (\ie~face, cranium, ears, eyes) that partly overlap with each other and capture statistical variations which are extremely difficult to incorporate into one single universal morphable model. Moreover, despite the increasing level of detail in the 3D face reconstruction from in-the-wild images, mainly attributed to recent advancements in deep learning, non of the available methods in the literature deal with the human tongue which is important for speech dynamics and improves the realness of the oral cavity. Finally, there is limited work on 3D facial geometric enchantments and translations from different capturing systems due to extremely limited availability of 3D dasasets tailored for this task. This thesis aims at tackling these shortcomings in all four domains. A novel approach on how to combine and enrich existing 3DMMs without the underline raw data is proposed. We introduce two methods for solving this problem: i. use a regressor to complete missing parts of one model using the other, ii. use a Gaussian Process framework to blend covariance matrices from multiple models. We show case our approach by combining existing face and head 3DMMs with different templates and statistical variations. Furthermore, we introduce to the research community the first Universal Head Model (UHM) which holds important statistical variation across all key structures of the human head that have an important contribution to to the appearance and identity of a person. We later show case how this model is used to create full head appearances from single in-the-wild images, thus making significant improvements toward the step of realist human head digitization from data-deficient sources. Additionally, we present the first method that accurately reconstructs the human tongue from single images by utilizing a novel generative framework which models directly the highly deformable surface of the human tongue and seamlessly merges it with our universal head model for more realist representations of the oral cavity dynamics. Lastly, in this thesis, it is presented a novel generative pipeline capable of converting and enhancing low to high quality 3D facial scans. This will potentially aid depth sensor applications by increasing the quality of the output data while maintaining a low cost. It is also shown that the proposed framework can be extended to handle translations between various expressions on demand.
- Published
- 2021
- Full Text
- View/download PDF
13. Towards a complete 3D morphable model of the human head
- Author
-
Ploumpis, Stylianos, Ververas, Evangelos, Sullivan, Eimear O', Moschoglou, Stylianos, Wang, Haoyang, Pears, Nick, Smith, William A. P., Gecer, Baris, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Three-dimensional Morphable Models (3DMMs) are powerful statistical tools for representing the 3D shapes and textures of an object class. Here we present the most complete 3DMM of the human head to date that includes face, cranium, ears, eyes, teeth and tongue. To achieve this, we propose two methods for combining existing 3DMMs of different overlapping head parts: i. use a regressor to complete missing parts of one model using the other, ii. use the Gaussian Process framework to blend covariance matrices from multiple models. Thus we build a new combined face-and-head shape model that blends the variability and facial detail of an existing face model (the LSFM) with the full head modelling capability of an existing head model (the LYHM). Then we construct and fuse a highly-detailed ear model to extend the variation of the ear shape. Eye and eye region models are incorporated into the head model, along with basic models of the teeth, tongue and inner mouth cavity. The new model achieves state-of-the-art performance. We use our model to reconstruct full head representations from single, unconstrained images allowing us to parameterize craniofacial shape and texture, along with the ear shape, eye gaze and eye color., Comment: 18 pages, 18 figures, submitted to Transactions on Pattern Analysis and Machine Intelligence (TPAMI) on the 9th of October as an extension paper of the original oral CVPR paper : arXiv:1903.03785
- Published
- 2019
14. Synthesizing Coupled 3D Face Modalities by Trunk-Branch Generative Adversarial Networks
- Author
-
Gecer, Baris, Lattas, Alexander, Ploumpis, Stylianos, Deng, Jiankang, Papaioannou, Athanasios, Moschoglou, Stylianos, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Graphics - Abstract
Generating realistic 3D faces is of high importance for computer graphics and computer vision applications. Generally, research on 3D face generation revolves around linear statistical models of the facial surface. Nevertheless, these models cannot represent faithfully either the facial texture or the normals of the face, which are very crucial for photo-realistic face synthesis. Recently, it was demonstrated that Generative Adversarial Networks (GANs) can be used for generating high-quality textures of faces. Nevertheless, the generation process either omits the geometry and normals, or independent processes are used to produce 3D shape information. In this paper, we present the first methodology that generates high-quality texture, shape, and normals jointly, which can be used for photo-realistic synthesis. To do so, we propose a novel GAN that can generate data from different modalities while exploiting their correlations. Furthermore, we demonstrate how we can condition the generation on the expression and create faces with various facial expressions. The qualitative results shown in this paper are compressed due to size limitations, full-resolution results and the accompanying video can be found in the supplementary documents. The code and models are available at the project page: https://github.com/barisgecer/TBGAN., Comment: Check project page: https://github.com/barisgecer/TBGAN for the full resolution results and the accompanying video
- Published
- 2019
- Full Text
- View/download PDF
15. Neural 3D Morphable Models: Spiral Convolutional Networks for 3D Shape Representation Learning and Generation
- Author
-
Bouritsas, Giorgos, Bokhnyak, Sergiy, Ploumpis, Stylianos, Bronstein, Michael, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence ,Computer Science - Graphics ,Computer Science - Machine Learning - Abstract
Generative models for 3D geometric data arise in many important applications in 3D computer vision and graphics. In this paper, we focus on 3D deformable shapes that share a common topological structure, such as human faces and bodies. Morphable Models and their variants, despite their linear formulation, have been widely used for shape representation, while most of the recently proposed nonlinear approaches resort to intermediate representations, such as 3D voxel grids or 2D views. In this work, we introduce a novel graph convolutional operator, acting directly on the 3D mesh, that explicitly models the inductive bias of the fixed underlying graph. This is achieved by enforcing consistent local orderings of the vertices of the graph, through the spiral operator, thus breaking the permutation invariance property that is adopted by all the prior work on Graph Neural Networks. Our operator comes by construction with desirable properties (anisotropic, topology-aware, lightweight, easy-to-optimise), and by using it as a building block for traditional deep generative architectures, we demonstrate state-of-the-art results on a variety of 3D shape datasets compared to the linear Morphable Model and other graph convolutional operators., Comment: to appear at ICCV 2019
- Published
- 2019
- Full Text
- View/download PDF
16. 3DFaceGAN: Adversarial Nets for 3D Face Representation, Generation, and Translation
- Author
-
Moschoglou, Stylianos, Ploumpis, Stylianos, Nicolaou, Mihalis, Papaioannou, Athanasios, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Over the past few years, Generative Adversarial Networks (GANs) have garnered increased interest among researchers in Computer Vision, with applications including, but not limited to, image generation, translation, imputation, and super-resolution. Nevertheless, no GAN-based method has been proposed in the literature that can successfully represent, generate or translate 3D facial shapes (meshes). This can be primarily attributed to two facts, namely that (a) publicly available 3D face databases are scarce as well as limited in terms of sample size and variability (e.g., few subjects, little diversity in race and gender), and (b) mesh convolutions for deep networks present several challenges that are not entirely tackled in the literature, leading to operator approximations and model instability, often failing to preserve high-frequency components of the distribution. As a result, linear methods such as Principal Component Analysis (PCA) have been mainly utilized towards 3D shape analysis, despite being unable to capture non-linearities and high frequency details of the 3D face - such as eyelid and lip variations. In this work, we present 3DFaceGAN, the first GAN tailored towards modeling the distribution of 3D facial surfaces, while retaining the high frequency details of 3D face shapes. We conduct an extensive series of both qualitative and quantitative experiments, where the merits of 3DFaceGAN are clearly demonstrated against other, state-of-the-art methods in tasks such as 3D shape representation, generation, and translation., Comment: 15 pages, 12 figures. Submitted to International Journal of Computer Vision (IJCV), special issue: Generative Adversarial Networks for Computer Vision
- Published
- 2019
17. Combining 3D Morphable Models: A Large scale Face-and-Head Model
- Author
-
Ploumpis, Stylianos, Wang, Haoyang, Pears, Nick, Smith, William A. P., and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Three-dimensional Morphable Models (3DMMs) are powerful statistical tools for representing the 3D surfaces of an object class. In this context, we identify an interesting question that has previously not received research attention: is it possible to combine two or more 3DMMs that (a) are built using different templates that perhaps only partly overlap, (b) have different representation capabilities and (c) are built from different datasets that may not be publicly-available? In answering this question, we make two contributions. First, we propose two methods for solving this problem: i. use a regressor to complete missing parts of one model using the other, ii. use the Gaussian Process framework to blend covariance matrices from multiple models. Second, as an example application of our approach, we build a new face-and-head shape model that combines the variability and facial detail of the LSFM with the full head modelling of the LYHM. The resulting combined shape model achieves state-of-the-art performance and outperforms existing head models by a large margin. Finally, as an application experiment, we reconstruct full head representations from single, unconstrained images by utilizing our proposed large-scale model in conjunction with the FaceWarehouse blendshapes for handling expressions., Comment: 9 pages, 8 figures. To appear in the Proceedings of Computer Vision and Pattern Recognition (CVPR), June 2019, Los Angeles, USA
- Published
- 2019
18. GANFIT: Generative Adversarial Network Fitting for High Fidelity 3D Face Reconstruction
- Author
-
Gecer, Baris, Ploumpis, Stylianos, Kotsia, Irene, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
In the past few years, a lot of work has been done towards reconstructing the 3D facial structure from single images by capitalizing on the power of Deep Convolutional Neural Networks (DCNNs). In the most recent works, differentiable renderers were employed in order to learn the relationship between the facial identity features and the parameters of a 3D morphable model for shape and texture. The texture features either correspond to components of a linear texture space or are learned by auto-encoders directly from in-the-wild images. In all cases, the quality of the facial texture reconstruction of the state-of-the-art methods is still not capable of modeling textures in high fidelity. In this paper, we take a radically different approach and harness the power of Generative Adversarial Networks (GANs) and DCNNs in order to reconstruct the facial texture and shape from single images. That is, we utilize GANs to train a very powerful generator of facial texture in UV space. Then, we revisit the original 3D Morphable Models (3DMMs) fitting approaches making use of non-linear optimization to find the optimal latent parameters that best reconstruct the test image but under a new perspective. We optimize the parameters with the supervision of pretrained deep identity features through our end-to-end differentiable framework. We demonstrate excellent results in photorealistic and identity preserving 3D face reconstructions and achieve for the first time, to the best of our knowledge, facial texture reconstruction with high-frequency details., Comment: CVPR 2019 camera ready; Check project page: https://github.com/barisgecer/ganfit for full resolution results and more
- Published
- 2019
- Full Text
- View/download PDF
19. MimicME: A Large Scale Diverse 4D Database for Facial Expression Analysis
- Author
-
Papaioannou, Athanasios, Gecer, Baris, Cheng, Shiyang, Chrysos, Grigorios, Deng, Jiankang, Fotiadou, Eftychia, Kampouris, Christos, Kollias, Dimitrios, Moschoglou, Stylianos, Songsri-In, Kritaphat, Ploumpis, Stylianos, Trigeorgis, George, Tzirakis, Panagiotis, Ververas, Evangelos, Zhou, Yuxiang, Ponniah, Allan, Roussos, Anastasios, Zafeiriou, Stefanos, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Avidan, Shai, editor, Brostow, Gabriel, editor, Cissé, Moustapha, editor, Farinella, Giovanni Maria, editor, and Hassner, Tal, editor
- Published
- 2022
- Full Text
- View/download PDF
20. A scalable mass customisation design process for 3D-printed respirator mask to combat COVID-19
- Author
-
Li, Shiya, Waheed, Usman, Bahshwan, Mohanad, Wang, Louis Zizhao, Kalossaka, Livia Mariadaria, Choi, Jiwoo, Kundrak, Franciska, Lattas, Alexandros, Ploumpis, Stylianos, Zafeiriou, Stefanos, and Myant, Connor William
- Published
- 2021
- Full Text
- View/download PDF
21. 3D Face Morphable Models 'In-the-Wild'
- Author
-
Booth, James, Antonakos, Epameinondas, Ploumpis, Stylianos, Trigeorgis, George, Panagakis, Yannis, and Zafeiriou, Stefanos
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
3D Morphable Models (3DMMs) are powerful statistical models of 3D facial shape and texture, and among the state-of-the-art methods for reconstructing facial shape from single images. With the advent of new 3D sensors, many 3D facial datasets have been collected containing both neutral as well as expressive faces. However, all datasets are captured under controlled conditions. Thus, even though powerful 3D facial shape models can be learnt from such data, it is difficult to build statistical texture models that are sufficient to reconstruct faces captured in unconstrained conditions ("in-the-wild"). In this paper, we propose the first, to the best of our knowledge, "in-the-wild" 3DMM by combining a powerful statistical model of facial shape, which describes both identity and expression, with an "in-the-wild" texture model. We show that the employment of such an "in-the-wild" texture model greatly simplifies the fitting procedure, because there is no need to optimize with regards to the illumination parameters. Furthermore, we propose a new fast algorithm for fitting the 3DMM in arbitrary images. Finally, we have captured the first 3D facial database with relatively unconstrained conditions and report quantitative evaluations with state-of-the-art performance. Complementary qualitative reconstruction results are demonstrated on standard "in-the-wild" facial databases. An open source implementation of our technique is released as part of the Menpo Project.
- Published
- 2017
22. Synthesizing Coupled 3D Face Modalities by Trunk-Branch Generative Adversarial Networks
- Author
-
Gecer, Baris, Lattas, Alexandros, Ploumpis, Stylianos, Deng, Jiankang, Papaioannou, Athanasios, Moschoglou, Stylianos, Zafeiriou, Stefanos, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Vedaldi, Andrea, editor, Bischof, Horst, editor, Brox, Thomas, editor, and Frahm, Jan-Michael, editor
- Published
- 2020
- Full Text
- View/download PDF
23. 3DFaceGAN: Adversarial Nets for 3D Face Representation, Generation, and Translation
- Author
-
Moschoglou, Stylianos, Ploumpis, Stylianos, Nicolaou, Mihalis A., Papaioannou, Athanasios, and Zafeiriou, Stefanos
- Published
- 2020
- Full Text
- View/download PDF
24. Multi-Attribute Probabilistic Linear Discriminant Analysis for 3D Facial Shapes
- Author
-
Moschoglou, Stylianos, Ploumpis, Stylianos, Nicolaou, Mihalis A., Zafeiriou, Stefanos, Hutchison, David, Editorial Board Member, Kanade, Takeo, Editorial Board Member, Kittler, Josef, Editorial Board Member, Kleinberg, Jon M., Editorial Board Member, Mattern, Friedemann, Editorial Board Member, Mitchell, John C., Editorial Board Member, Naor, Moni, Editorial Board Member, Pandu Rangan, C., Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Terzopoulos, Demetri, Editorial Board Member, Tygar, Doug, Editorial Board Member, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Jawahar, C. V., editor, Li, Hongdong, editor, Mori, Greg, editor, and Schindler, Konrad, editor
- Published
- 2019
- Full Text
- View/download PDF
25. ViPED: On-road vehicle passenger detection for autonomous vehicles
- Author
-
Amanatiadis, Angelos, Karakasis, Evangelos, Bampis, Loukas, Ploumpis, Stylianos, and Gasteratos, Antonios
- Published
- 2019
- Full Text
- View/download PDF
26. Synthesizing Coupled 3D Face Modalities by Trunk-Branch Generative Adversarial Networks
- Author
-
Gecer, Baris, primary, Lattas, Alexandros, additional, Ploumpis, Stylianos, additional, Deng, Jiankang, additional, Papaioannou, Athanasios, additional, Moschoglou, Stylianos, additional, and Zafeiriou, Stefanos, additional
- Published
- 2020
- Full Text
- View/download PDF
27. Learning to Generate Customized Dynamic 3D Facial Expressions
- Author
-
Potamias, Rolandos Alexandros, primary, Zheng, Jiali, additional, Ploumpis, Stylianos, additional, Bouritsas, Giorgos, additional, Ververas, Evangelos, additional, and Zafeiriou, Stefanos, additional
- Published
- 2020
- Full Text
- View/download PDF
28. Handy: Towards a High Fidelity 3D Hand Shape and Appearance Model
- Author
-
Potamias, Rolandos Alexandros, primary, Ploumpis, Stylianos, additional, Moschoglou, Stylianos, additional, Triantafyllou, Vasileios, additional, and Zafeiriou, Stefanos, additional
- Published
- 2023
- Full Text
- View/download PDF
29. FitMe: Deep Photorealistic 3D Morphable Model Avatars
- Author
-
Lattas, Alexandros, primary, Moschoglou, Stylianos, additional, Ploumpis, Stylianos, additional, Gecer, Baris, additional, Deng, Jiankang, additional, and Zafeiriou, Stefanos, additional
- Published
- 2023
- Full Text
- View/download PDF
30. Multi-Attribute Probabilistic Linear Discriminant Analysis for 3D Facial Shapes
- Author
-
Moschoglou, Stylianos, primary, Ploumpis, Stylianos, additional, Nicolaou, Mihalis A., additional, and Zafeiriou, Stefanos, additional
- Published
- 2019
- Full Text
- View/download PDF
31. A stereo matching approach based on particle filters and scattered control landmarks
- Author
-
Ploumpis, Stylianos, Amanatiadis, Angelos, and Gasteratos, Antonios
- Published
- 2015
- Full Text
- View/download PDF
32. Dynamic Neural Portraits
- Author
-
Doukas, Michail Christos, Ploumpis, Stylianos, and Zafeiriou, Stefanos
- Subjects
FOS: Computer and information sciences ,Computer Vision and Pattern Recognition (cs.CV) ,Computer Science - Computer Vision and Pattern Recognition - Abstract
We present Dynamic Neural Portraits, a novel approach to the problem of full-head reenactment. Our method generates photo-realistic video portraits by explicitly controlling head pose, facial expressions and eye gaze. Our proposed architecture is different from existing methods that rely on GAN-based image-to-image translation networks for transforming renderings of 3D faces into photo-realistic images. Instead, we build our system upon a 2D coordinate-based MLP with controllable dynamics. Our intuition to adopt a 2D-based representation, as opposed to recent 3D NeRF-like systems, stems from the fact that video portraits are captured by monocular stationary cameras, therefore, only a single viewpoint of the scene is available. Primarily, we condition our generative model on expression blendshapes, nonetheless, we show that our system can be successfully driven by audio features as well. Our experiments demonstrate that the proposed method is 270 times faster than recent NeRF-based reenactment methods, with our networks achieving speeds of 24 fps for resolutions up to 1024 x 1024, while outperforming prior works in terms of visual quality., In IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) 2023
- Published
- 2023
- Full Text
- View/download PDF
33. AvatarMe++: Facial Shape and BRDF Inference With Photorealistic Rendering-Aware GANs
- Author
-
Lattas, Alexandros, primary, Moschoglou, Stylianos, additional, Ploumpis, Stylianos, additional, Gecer, Baris, additional, Ghosh, Abhijeet, additional, and Zafeiriou, Stefanos, additional
- Published
- 2022
- Full Text
- View/download PDF
34. 3D human tongue reconstruction from single 'in-the-wild' images
- Author
-
Ploumpis, Stylianos, Moschoglou, Stylianos, Triantafyllou, Vasileios, and Zafeiriou, Stefanos
- Subjects
FOS: Computer and information sciences ,Computer Science - Graphics ,Artificial Intelligence (cs.AI) ,Computer Science - Artificial Intelligence ,Computer Vision and Pattern Recognition (cs.CV) ,Computer Science - Computer Vision and Pattern Recognition ,Graphics (cs.GR) - Abstract
3D face reconstruction from a single image is a task that has garnered increased interest in the Computer Vision community, especially due to its broad use in a number of applications such as realistic 3D avatar creation, pose invariant face recognition and face hallucination. Since the introduction of the 3D Morphable Model in the late 90's, we witnessed an explosion of research aiming at particularly tackling this task. Nevertheless, despite the increasing level of detail in the 3D face reconstructions from single images mainly attributed to deep learning advances, finer and highly deformable components of the face such as the tongue are still absent from all 3D face models in the literature, although being very important for the realness of the 3D avatar representations. In this work we present the first, to the best of our knowledge, end-to-end trainable pipeline that accurately reconstructs the 3D face together with the tongue. Moreover, we make this pipeline robust in "in-the-wild" images by introducing a novel GAN method tailored for 3D tongue surface generation. Finally, we make publicly available to the community the first diverse tongue dataset, consisting of 1,800 raw scans of 700 individuals varying in gender, age, and ethnicity backgrounds. As we demonstrate in an extensive series of quantitative as well as qualitative experiments, our model proves to be robust and realistically captures the 3D tongue structure, even in adverse "in-the-wild" conditions., Comment: 10 pages, 9 figures
- Published
- 2022
- Full Text
- View/download PDF
35. Neural Mesh Simplification
- Author
-
Potamias, Rolandos Alexandros, primary, Ploumpis, Stylianos, additional, and Zafeiriou, Stefanos, additional
- Published
- 2022
- Full Text
- View/download PDF
36. Towards a Complete 3D Morphable Model of the Human Head
- Author
-
Ploumpis, Stylianos, primary, Ververas, Evangelos, additional, Sullivan, Eimear Oa, additional, Moschoglou, Stylianos, additional, Wang, Haoyang, additional, Pears, Nick, additional, Smith, William A. P., additional, Gecer, Baris, additional, and Zafeiriou, Stefanos, additional
- Published
- 2021
- Full Text
- View/download PDF
37. AvatarMe ++ : Facial Shape and BRDF Inference With Photorealistic Rendering-Aware GANs.
- Author
-
Lattas, Alexandros, Moschoglou, Stylianos, Ploumpis, Stylianos, Gecer, Baris, Ghosh, Abhijeet, and Zafeiriou, Stefanos
- Subjects
GENERATIVE adversarial networks ,FACE ,TASK analysis - Abstract
Over the last years, with the advent of Generative Adversarial Networks (GANs), many face analysis tasks have accomplished astounding performance, with applications including, but not limited to, face generation and 3D face reconstruction from a single “in-the-wild” image. Nevertheless, to the best of our knowledge, there is no method which can produce render-ready high-resolution 3D faces from “in-the-wild” images and this can be attributed to the: (a) scarcity of available data for training, and (b) lack of robust methodologies that can successfully be applied on very high-resolution data. In this paper, we introduce the first method that is able to reconstruct photorealistic render-ready 3D facial geometry and BRDF from a single “in-the-wild” image. To achieve this, we capture a large dataset of facial shape and reflectance, which we have made public. Moreover, we define a fast and photorealistic differentiable rendering methodology with accurate facial skin diffuse and specular reflection, self-occlusion and subsurface scattering approximation. With this, we train a network that disentangles the facial diffuse and specular reflectance components from a mesh and texture with baked illumination, scanned or reconstructed with a 3DMM fitting method. As we demonstrate in a series of qualitative and quantitative experiments, our method outperforms the existing arts by a significant margin and reconstructs authentic, 4K by 6K-resolution 3D faces from a single low-resolution image, that are ready to be rendered in various applications and bridge the uncanny valley. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
38. Fast-GANFIT: Generative Adversarial Network for High Fidelity 3D Face Reconstruction.
- Author
-
Gecer, Baris, Ploumpis, Stylianos, Kotsia, Irene, and Zafeiriou, Stefanos
- Subjects
- *
GENERATIVE adversarial networks , *ARTIFICIAL neural networks , *CONVOLUTIONAL neural networks , *FACE , *VECTOR spaces - Abstract
A lot of work has been done towards reconstructing the 3D facial structure from single images by capitalizing on the power of deep convolutional neural networks (DCNNs). In the recent works, the texture features either correspond to components of a linear texture space or are learned by auto-encoders directly from in-the-wild images. In all cases, the quality of the facial texture reconstruction is still not capable of modeling facial texture with high-frequency details. In this paper, we take a radically different approach and harness the power of generative adversarial networks (GANs) and DCNNs in order to reconstruct the facial texture and shape from single images. That is, we utilize GANs to train a very powerful facial texture prior from a large-scale 3D texture dataset. Then, we revisit the original 3D Morphable Models (3DMMs) fitting making use of non-linear optimization to find the optimal latent parameters that best reconstruct the test image but under a new perspective. In order to be robust towards initialisation and expedite the fitting process, we propose a novel self-supervised regression based approach. We demonstrate excellent results in photorealistic and identity preserving 3D face reconstructions and achieve for the first time, to the best of our knowledge, facial texture reconstruction with high-frequency details. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
39. Fast-GANFIT: Generative Adversarial Network for High Fidelity 3D Face Reconstruction
- Author
-
Gecer, Baris, primary, Ploumpis, Stylianos, additional, Kotsia, Irene, additional, and Zafeiriou, Stefanos P, additional
- Published
- 2021
- Full Text
- View/download PDF
40. Design Automation for Mass Customisation via Additive Manufacture: A Case Study on Continuous Positive Airway Pressure Mask
- Author
-
Li, Shiya, additional, Ploumpis, Stylianos, additional, Zafeiriou, Stefanos, additional, and Myant, Connor, additional
- Published
- 2020
- Full Text
- View/download PDF
41. AvatarMe: Realistically Renderable 3D Facial Reconstruction “In-the-Wild”
- Author
-
Lattas, Alexandros, primary, Moschoglou, Stylianos, additional, Gecer, Baris, additional, Ploumpis, Stylianos, additional, Triantafyllou, Vasileios, additional, Ghosh, Abhijeet, additional, and Zafeiriou, Stefanos, additional
- Published
- 2020
- Full Text
- View/download PDF
42. 3D head morphable models and beyond: algorithms and applications
- Author
-
Ploumpis, Stylianos A. and Zafeiriou, Stefanos
- Abstract
It has been more than 20 year since the introduction of 3D morphable models (3DMM) in the computer vision literature. They were proposed as a face representation based on principal components analysis for the task of image analysis, photorealist-manipulation, and 3D reconstruction from single images. Even so, to this date, the applications of such models are limited by a number of factors. Firstly, training correctly 3DMMs require a vast amount of 3D data that most of the times are not publicly available to the research community due to increasingly stringent data protection regulations. Hence, it is extremely difficult to combine and enrich multiple attributes of the human face/head without the initial 3D images. Additionally, many 3DMMs utilize different templates that describe distinct parts of the human face/head (\ie~face, cranium, ears, eyes) that partly overlap with each other and capture statistical variations which are extremely difficult to incorporate into one single universal morphable model. Moreover, despite the increasing level of detail in the 3D face reconstruction from in-the-wild images, mainly attributed to recent advancements in deep learning, non of the available methods in the literature deal with the human tongue which is important for speech dynamics and improves the realness of the oral cavity. Finally, there is limited work on 3D facial geometric enchantments and translations from different capturing systems due to extremely limited availability of 3D dasasets tailored for this task. This thesis aims at tackling these shortcomings in all four domains. A novel approach on how to combine and enrich existing 3DMMs without the underline raw data is proposed. We introduce two methods for solving this problem: i. use a regressor to complete missing parts of one model using the other, ii. use a Gaussian Process framework to blend covariance matrices from multiple models. We show case our approach by combining existing face and head 3DMMs with different templates and statistical variations. Furthermore, we introduce to the research community the first Universal Head Model (UHM) which holds important statistical variation across all key structures of the human head that have an important contribution to to the appearance and identity of a person. We later show case how this model is used to create full head appearances from single in-the-wild images, thus making significant improvements toward the step of realist human head digitization from data-deficient sources. Additionally, we present the first method that accurately reconstructs the human tongue from single images by utilizing a novel generative framework which models directly the highly deformable surface of the human tongue and seamlessly merges it with our universal head model for more realist representations of the oral cavity dynamics. Lastly, in this thesis, it is presented a novel generative pipeline capable of converting and enhancing low to high quality 3D facial scans. This will potentially aid depth sensor applications by increasing the quality of the output data while maintaining a low cost. It is also shown that the proposed framework can be extended to handle translations between various expressions on demand. Open Access
- Published
- 2020
- Full Text
- View/download PDF
43. Neural 3D Morphable Models: Spiral Convolutional Networks for 3D Shape Representation Learning and Generation
- Author
-
Bouritsas, Giorgos, primary, Bokhnyak, Sergiy, additional, Ploumpis, Stylianos, additional, Zafeiriou, Stefanos, additional, and Bronstein, Michael, additional
- Published
- 2019
- Full Text
- View/download PDF
44. GANFIT: Generative Adversarial Network Fitting for High Fidelity 3D Face Reconstruction
- Author
-
Gecer, Baris, primary, Ploumpis, Stylianos, additional, Kotsia, Irene, additional, and Zafeiriou, Stefanos, additional
- Published
- 2019
- Full Text
- View/download PDF
45. Combining 3D Morphable Models: A Large Scale Face-And-Head Model
- Author
-
Ploumpis, Stylianos, primary, Wang, Haoyang, additional, Pears, Nick, additional, Smith, William A. P., additional, and Zafeiriou, Stefanos, additional
- Published
- 2019
- Full Text
- View/download PDF
46. 3D Reconstruction of “In-the-Wild” Faces in Images and Videos
- Author
-
Booth, James, primary, Roussos, Anastasios, additional, Ververas, Evangelos, additional, Antonakos, Epameinondas, additional, Ploumpis, Stylianos, additional, Panagakis, Yannis, additional, and Zafeiriou, Stefanos, additional
- Published
- 2018
- Full Text
- View/download PDF
47. 3D Face Morphable Models "In-the-Wild"
- Author
-
Booth, James, primary, Antonakos, Epameinondas, additional, Ploumpis, Stylianos, additional, Trigeorgis, George, additional, Panagakis, Yannis, additional, and Zafeiriou, Stefanos, additional
- Published
- 2017
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.