16 results on '"Tewari, Ayush"'
Search Results
2. Characterization of tropical plane curves up to genus six
- Author
-
Tewari, Ayush Kumar
- Published
- 2023
- Full Text
- View/download PDF
3. Point line geometry in the tropical plane
- Author
-
Tewari, Ayush Kumar
- Published
- 2023
- Full Text
- View/download PDF
4. Moduli dimensions of lattice polygons
- Author
-
Echavarria, Marino, Everett, Max, Huang, Shinyu, Jacoby, Liza, Morrison, Ralph, Tewari, Ayush K., Vlad, Raluca, and Weber, Ben
- Published
- 2022
- Full Text
- View/download PDF
5. Forbidden patterns in tropical plane curves
- Author
-
Joswig, Michael and Tewari, Ayush Kumar
- Published
- 2021
- Full Text
- View/download PDF
6. AvatarStudio: Text-Driven Editing of 3D Dynamic Human Head Avatars.
- Author
-
Mendiratta, Mohit, Pan, Xingang, Elgharib, Mohamed, Teotia, Kartik, R, Mallikarjun B, Tewari, Ayush, Golyanik, Vladislav, Kortylewski, Adam, and Theobalt, Christian
- Abstract
Capturing and editing full-head performances enables the creation of virtual characters with various applications such as extended reality and media production. The past few years witnessed a steep rise in the photorealism of human head avatars. Such avatars can be controlled through different input data modalities, including RGB, audio, depth, IMUs, and others. While these data modalities provide effective means of control, they mostly focus on editing the head movements such as the facial expressions, head pose, and/or camera viewpoint. In this paper, we propose AvatarStudio, a text-based method for editing the appearance of a dynamic full head avatar. Our approach builds on existing work to capture dynamic performances of human heads using Neural Radiance Field (NeRF) and edits this representation with a text-to-image diffusion model. Specifically, we introduce an optimization strategy for incorporating multiple keyframes representing different camera viewpoints and time stamps of a video performance into a single diffusion model. Using this personalized diffusion model, we edit the dynamic NeRF by introducing view-and-time-aware Score Distillation Sampling (VT-SDS) following a model-based guidance approach. Our method edits the full head in a canonical space and then propagates these edits to the remaining time steps via a pre-trained deformation network. We evaluate our method visually and numerically via a user study, and results show that our method outperforms existing approaches. Our experiments validate the design choices of our method and highlight that our edits are genuine, personalized, as well as 3D- and time-consistent. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
7. Diffusion Posterior Illumination for Ambiguity-Aware Inverse Rendering.
- Author
-
Lyu, Linjie, Tewari, Ayush, Habermann, Marc, Saito, Shunsuke, Zollhöfer, Michael, Leimkühler, Thomas, and Theobalt, Christian
- Abstract
Inverse rendering, the process of inferring scene properties from images, is a challenging inverse problem. The task is ill-posed, as many different scene configurations can give rise to the same image. Most existing solutions incorporate priors into the inverse-rendering pipeline to encourage plausible solutions, but they do not consider the inherent ambiguities and the multi-modal distribution of possible decompositions. In this work, we propose a novel scheme that integrates a denoising diffusion probabilistic model pre-trained on natural illumination maps into an optimization framework involving a differentiable path tracer. The proposed method allows sampling from combinations of illumination and spatially-varying surface materials that are, both, natural and explain the image observations. We further conduct an extensive comparative study of different priors on illumination used in previous work on inverse rendering. Our method excels in recovering materials and producing highly realistic and diverse environment map samples that faithfully explain the illumination of the input images. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
8. ModalNeRF: Neural Modal Analysis and Synthesis for Free‐Viewpoint Navigation in Dynamically Vibrating Scenes.
- Author
-
Petitjean, Automne, Poirier‐Ginter, Yohan, Tewari, Ayush, Cordonnier, Guillaume, and Drettakis, George
- Subjects
PERIODIC motion ,MOTION capture (Human mechanics) ,GENOME editing ,CAMCORDERS ,MOTION analysis ,MOTION ,MODAL analysis ,SYNTHETIC biology - Abstract
Recent advances in Neural Radiance Fields enable the capture of scenes with motion. However, editing the motion is hard; no existing method allows editing beyond the space of motion existing in the original video, nor editing based on physics. We present the first approach that allows physically‐based editing of motion in a scene captured with a single hand‐held video camera, containing vibrating or periodic motion. We first introduce a Lagrangian representation, representing motion as the displacement of particles, which is learned while training a radiance field. We use these particles to create a continuous representation of motion over the sequence, which is then used to perform a modal analysis of the motion thanks to a Fourier transform on the particle displacement over time. The resulting extracted modes allow motion synthesis, and easy editing of the motion, while inheriting the ability for free‐viewpoint synthesis in the captured 3D scene from the radiance field. We demonstrate our new method on synthetic and real captured scenes. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
9. PhotoApp: photorealistic appearance editing of head portraits.
- Author
-
R, Mallikarjun B, Tewari, Ayush, Dib, Abdallah, Weyrich, Tim, Bickel, Bernd, Seidel, Hans-Peter, Pfister, Hanspeter, Matusik, Wojciech, Chevallier, Louis, Elgharib, Mohamed, and Theobalt, Christian
- Subjects
SUPERVISED learning ,EDITING ,GENERALIZATION - Abstract
Photorealistic editing of head portraits is a challenging task as humans are very sensitive to inconsistencies in faces. We present an approach for high-quality intuitive editing of the camera viewpoint and scene illumination (parameterised with an environment map) in a portrait image. This requires our method to capture and control the full reflectance field of the person in the image. Most editing approaches rely on supervised learning using training data captured with setups such as light and camera stages. Such datasets are expensive to acquire, not readily available and do not capture all the rich variations of in-the-wild portrait images. In addition, most supervised approaches only focus on relighting, and do not allow camera viewpoint editing. Thus, they only capture and control a subset of the reflectance field. Recently, portrait editing has been demonstrated by operating in the generative model space of StyleGAN. While such approaches do not require direct supervision, there is a significant loss of quality when compared to the supervised approaches. In this paper, we present a method which learns from limited supervised training data. The training images only include people in a fixed neutral expression with eyes closed, without much hair or background variations. Each person is captured under 150 one-light-at-a-time conditions and under 8 camera poses. Instead of training directly in the image space, we design a supervised problem which learns transformations in the latent space of StyleGAN. This combines the best of supervised learning and generative adversarial modeling. We show that the StyleGAN prior allows for generalisation to different expressions, hairstyles and backgrounds. This produces high-quality photorealistic results for in-the-wild images and significantly outperforms existing methods. Our approach can edit the illumination and pose simultaneously, and runs at interactive rates. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
10. Egocentric videoconferencing.
- Author
-
Elgharib, Mohamed, Mendiratta, Mohit, Thies, Justus, Niessner, Matthias, Seidel, Hans-Peter, Tewari, Ayush, Golyanik, Vladislav, and Theobalt, Christian
- Subjects
VIDEOCONFERENCING ,NONVERBAL communication ,FACIAL expression ,BLINKING (Physiology) ,CAMERA phones ,BINOCULAR vision ,EGOCENTRIC bias - Abstract
We introduce a method for egocentric videoconferencing that enables hands-free video calls, for instance by people wearing smart glasses or other mixed-reality devices. Videoconferencing portrays valuable non-verbal communication and face expression cues, but usually requires a front-facing camera. Using a frontal camera in a hands-free setting when a person is on the move is impractical. Even holding a mobile phone camera in the front of the face while sitting for a long duration is not convenient. To overcome these issues, we propose a low-cost wearable egocentric camera setup that can be integrated into smart glasses. Our goal is to mimic a classical video call, and therefore, we transform the egocentric perspective of this camera into a front facing video. To this end, we employ a conditional generative adversarial neural network that learns a transition from the highly distorted egocentric views to frontal views common in videoconferencing. Our approach learns to transfer expression details directly from the egocentric view without using a complex intermediate parametric expressions model, as it is used by related face reenactment methods. We successfully handle subtle expressions, not easily captured by parametric blendshape-based solutions, e.g., tongue movement, eye movements, eye blinking, strong expressions and depth varying movements. To get control over the rigid head movements in the target view, we condition the generator on synthetic renderings of a moving neutral face. This allows us to synthesis results at different head poses. Our technique produces temporally smooth video-realistic renderings in real-time using a video-to-video translation network in conjunction with a temporal discriminator. We demonstrate the improved capabilities of our technique by comparing against related state-of-the art approaches. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
11. PIE: portrait image embedding for semantic control.
- Author
-
Tewari, Ayush, Elgharib, Mohamed, R, Mallikarjun B, Bernard, Florian, Seidel, Hans-Peter, Pérez, Patrick, Zollhöfer, Michael, and Theobalt, Christian
- Subjects
FACIAL expression ,NONLINEAR equations ,COMPUTER-generated imagery ,PORTRAITS ,AUTOMATION ,PIES - Abstract
Editing of portrait images is a very popular and important research topic with a large variety of applications. For ease of use, control should be provided via a semantically meaningful parameterization that is akin to computer animation controls. The vast majority of existing techniques do not provide such intuitive and fine-grained control, or only enable coarse editing of a single isolated control parameter. Very recently, high-quality semantically controlled editing has been demonstrated, however only on synthetically created StyleGAN images. We present the first approach for embedding real portrait images in the latent space of StyleGAN, which allows for intuitive editing of the head pose, facial expression, and scene illumination in the image. Semantic editing in parameter space is achieved based on StyleRig, a pretrained neural network that maps the control space of a 3D morphable face model to the latent space of the GAN. We design a novel hierarchical non-linear optimization problem to obtain the embedding. An identity preservation energy term allows spatially coherent edits while maintaining facial integrity. Our approach runs at interactive frame rates and thus allows the user to explore the space of possible edits. We evaluate our approach on a wide set of portrait photos, compare it to the current state of the art, and validate the effectiveness of its components in an ablation study. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
12. 3D Morphable Face Models—Past, Present, and Future.
- Author
-
Egger, Bernhard, Smith, William A. P., Tewari, Ayush, Wuhrer, Stefanie, Zollhoefer, Michael, Beeler, Thabo, Bernard, Florian, Bolkart, Timo, Kortylewski, Adam, Romdhani, Sami, Theobalt, Christian, Blanz, Volker, and Vetter, Thomas
- Subjects
IMAGE analysis ,COMPUTER vision ,COMPUTER graphics - Abstract
In this article, we provide a detailed survey of 3D Morphable Face Models over the 20 years since they were first proposed. The challenges in building and applying these models, namely, capture, modeling, image formation, and image analysis, are still active research topics, and we review the state-of-the-art in each of these areas. We also look ahead, identifying unsolved challenges, proposing directions for future research, and highlighting the broad range of current and future applications. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
13. High-Fidelity Monocular Face Reconstruction Based on an Unsupervised Model-Based Face Autoencoder.
- Author
-
Tewari, Ayush, Zollhofer, Michael, Bernard, Florian, Garrido, Pablo, Kim, Hyeongwoo, Perez, Patrick, and Theobalt, Christian
- Subjects
- *
SAMPLING (Process) , *IMAGE reconstruction - Abstract
In this work, we propose a novel model-based deep convolutional autoencoder that addresses the highly challenging problem of reconstructing a 3D human face from a single in-the-wild color image. To this end, we combine a convolutional encoder network with an expert-designed generative model that serves as decoder. The core innovation is the differentiable parametric decoder that encapsulates image formation analytically based on a generative model. Our decoder takes as input a code vector with exactly defined semantic meaning that encodes detailed face pose, shape, expression, skin reflectance, and scene illumination. Due to this new way of combining CNN-based with model-based face reconstruction, the CNN-based encoder learns to extract semantically meaningful parameters from a single monocular input image. For the first time, a CNN encoder and an expert-designed generative model can be trained end-to-end in an unsupervised manner, which renders training on very large (unlabeled) real world datasets feasible. The obtained reconstructions compare favorably to current state-of-the-art approaches in terms of quality and richness of representation. This work is an extended version of , where we additionally present a stochastic vertex sampling technique for faster training of our networks, and moreover, we propose and evaluate analysis-by-synthesis and shape-from-shading refinement approaches to achieve a high-fidelity reconstruction. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
14. Text-based editing of talking-head video.
- Author
-
Fried, Ohad, Tewari, Ayush, Zollhöfer, Michael, Finkelstein, Adam, Shechtman, Eli, Goldman, Dan B, Genova, Kyle, Jin, Zeyu, Theobalt, Christian, and Agrawala, Maneesh
- Subjects
EDITING ,SPEECH ,VIDEOS ,DUBBING ,RENDERING (Computer graphics) - Abstract
Editing talking-head video to change the speech content or to remove filler words is challenging. We propose a novel method to edit talking-head video based on its transcript to produce a realistic output video in which the dialogue of the speaker has been modified, while maintaining a seamless audio-visual flow (i.e. no jump cuts). Our method automatically annotates an input talking-head video with phonemes, visemes, 3D face pose and geometry, reflectance, expression and scene illumination per frame. To edit a video, the user has to only edit the transcript, and an optimization strategy then chooses segments of the input corpus as base material. The annotated parameters corresponding to the selected segments are seamlessly stitched together and used to produce an intermediate video representation in which the lower half of the face is rendered with a parametric face model. Finally, a recurrent video generation network transforms this representation to a photorealistic video that matches the edited transcript. We demonstrate a large variety of edits, such as the addition, removal, and alteration of words, as well as convincing language translation and full sentence synthesis. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
15. Deep video portraits.
- Author
-
Kim, Hyeongwoo, Carrido, Pablo, Tewari, Ayush, Xu, Weipeng, Thies, Justus, Niessner, Matthias, Pérez, Patrick, Richardt, Christian, Zollhöfer, Michael, and Theobalt, Christian
- Subjects
DEEP learning ,VIDEO recording ,COMPUTER-generated imagery ,THREE-dimensional imaging ,FACIAL expression - Abstract
We present a novel approach that enables photo-realistic re-animation of portrait videos using only an input video. In contrast to existing approaches that are restricted to manipulations of facial expressions only, we are the first to transfer the full 3D head position, head rotation, face expression, eye gaze, and eye blinking from a source actor to a portrait video of a target actor. The core of our approach is a generative neural network with a novel space-time architecture. The network takes as input synthetic renderings of a parametric face model, based on which it predicts photo-realistic video frames for a given target actor. The realism in this rendering-to-video transfer is achieved by careful adversarial training, and as a result, we can create modified target videos that mimic the behavior of the synthetically-created input. In order to enable source-to-target video re-animation, we render a synthetic target video with the reconstructed head animation parameters from a source video, and feed it into the trained network - thus taking full control of the target. With the ability to freely recombine source and target parameters, we are able to demonstrate a large variety of video rewrite applications without explicitly modeling hair, body or background. For instance, we can reenact the full head using interactive user-controlled editing, and realize high-fidelity visual dubbing. To demonstrate the high quality of our output, we conduct an extensive series of experiments and evaluations, where for instance a user study shows that our video edits are hard to detect. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
16. Convex lattice polygons with all lattice points visible.
- Author
-
Morrison, Ralph and Tewari, Ayush Kumar
- Subjects
- *
PLANE curves , *POLYGONS - Abstract
Two lattice points are visible to one another if there exist no other lattice points on the line segment connecting them. In this paper we study convex lattice polygons that contain a lattice point such that all other lattice points in the polygon are visible from it. We completely classify such polygons, show that there are finitely many of lattice width greater than 2, and computationally enumerate them. As an application of this classification, we prove new obstructions to graphs arising as skeleta of tropical plane curves. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.