|
Exploring Attribute Variations in Style-based GANs using Diffusion Models
Rishubh Parihar,
Prasanna Balaji,
Raghav Magazine,
Sarthak Vora,
Tejan Karmali,
Varun Jampani,
R. Venkatesh Babu
NeurIPS Workshop on Diffusion Models, 2023
arXiv
Existing attribute editing methods treat semantic attributes as binary, resulting in a single edit per attribute. However, attributes such
as eyeglasses, smiles, or hairstyles exhibit a vast range of diversity. In this work, we formulate the task of diverse attribute editing by
modeling the multidimensional nature of attribute edits. We capitalize on disentangled latent spaces of pretrained GANs and train a Denoising Diffusion
Probabilistic Model (DDPM) to learn the latent distribution for diverse edits. This leads to latent subspaces that enable diverse attribute editing.
Applying diffusion in the highly compressed latent space allows us to model rich distributions of edits within limited computational resources
|
|
Strata-NeRF: Neural Radiance Fields for Stratified Scenes
Ankit Dhiman,
R Srinath,
Harsh Rangwani,
Rishubh Parihar,
Lokesh R Boregowda,
Srinath Sridhar,
R. Venkatesh Babu
ICCV, 2023
project page
/
arXiv
/
code
Existing NeRF approaches concentrate on modelling a single object or a single level of a scene. However, in the real world, we may capture a
scene at multiple levels, resulting in a layered capture. For example, tourists usually capture a monument’s exterior structure before capturing the inner structure.
We propose Strata-NeRF, a single neural radiance field that implicitly captures a scene with multiple levels. Strata-NeRF achieves this by conditioning the
NeRFs on Vector Quantized (VQ) latent representations which allow sudden changes in scene structure.
|
|
We never go out of Style: Motion Disentanglement by Subspace Decomposition of Latent Space
Rishubh Parihar,
Raghav Magazine,
Piyush Tiwari,
R. Venkatesh Babu
CVPRW, AI for Content Creation Workshop, 2023
project page
/
arXiv
/
video
In this work, we propose a novel method to decompose motion in videos by using a pretrained image GAN model. We discover disentangled motion subspaces in the latent space of widely used style-based GAN models that
are semantically meaningful and control a single explainable motion component. The proposed method uses only a few (≈10) ground truth video sequences to obtain such subspaces.
|
|
Everything is in the Latent Space: Attribute Style Editing and Attribute Style Manipulation
by StyleGAN Latent Space Exploration
Rishubh Parihar,
Ankit Dhiman,
Tejan Karmali,
R. Venkatesh Babu
ACMMM, 2022
project page
/
arXiv
/
video
In this work we proposed a Few-shot Latent based Attribute Manipulation and Editing (FLAME) method, a simple yet effective framework to perform highly controlled image editing by latent space manipulation.
FLAME can generate highly realistic attribute edits and enables us to generate diverse attribute styles such as hair-styles, trained with only a few set of image pairs.
|
|
Hierarchical Semantic Regularization of Latent Spaces in StyleGANs
Tejan Karmali,
Rishubh Parihar,
Susmit Agrawal,
Harsh Rangwani,
Varun Jampani,
Manish Singh,
R. Venkatesh Babu
ECCV, 2022
project page
/
arXiv
Proposed a hierarchical regularizer during the training of StyleGAN models to induce smoothness properties in the W/W+ latent spaces.
The regularizer is implemented by guiding the intermediate features of the StyleGAN with the features from the pretrained feature extractors.
|
|
Spatio-Temporal Video Representation Learning via motion type classification
Rishubh Parihar,
Gaurav Ramola,
Ranajit Saha,
Ravi Kini,
Aniket Rege,
Sudha Velusamy
ICCVW - SRVU, 2021
arXiv
In this paper, we propose a novel approach for understanding object motions via motion type classification.
The proposed motion type classifier predicts a motion type for the video based on the trajectories of the objects present.
Our classifier assigns a motion type for the given video from the following five primitive motion classes: linear, projectile, oscillatory, local and random.
We demonstrate that the representations learned from the motion type classification generalizes well for the challenging downstream task of video retrieval.
Further, we proposed a recommendation system for video playback style based on the motion type classifier predictions.
|
|
FabSoften: Face Beautification via Dynamic Skin Smoothing, Guided
Feathering, and Texture Restoration
Sudha Velusamy,
Rishubh Parihar,
Ravi Kini,
Aniket Rege
CVPRW - NTIRE, 2020
pdf
We propose a real-time face softening approach that smooths blemishes in the facial skin region, followed by a wavelet band manipulation to restore
the underlying skin texture, which produces a highly appealing ‘beautified’ face that retains its natural appearance.
|
|
Scene Adaptive Cosmetic Makeup Transfer
Rishubh Parihar,
Aniket Dashpute,
Prem Kalra
Undergrduate Thesis
pdf
Given a source and a target image transferring makeup from the source image to the target image.
The transferred makeup should blend in the scene to provide natural look. To this end, we have developed
a complete framework which firstly relights the subject image to match the illumination of the target image.
We have generated 3D face models from single image and used them for realistic relighting. Following that
layer wise decomposition is done for relit source and target image and blending is done within corresponding layers
to transfer makeup. Finally we have additional modules in our framework to add facial accessories.
As we have generated 3D models of the source and target faces we were able to add accessories directly on 3D models
which resulted in natural looking output.
|
|
Image enhancement using pixel-transformer
Ankit Dhiman,
Rishubh Parihar
We empirically validate a generative model - PixelTrans- former, which infers distribution of the spatial signal given a sparse set of input samples
ex. image from the few ob- served pixels. We tested the method under two scenarios; when sampler polls from 1.) a noisy representation and 2.) a low-frequency representation of the underlying spatial signal. Also, we evaluated the model for the different number of encoder and decoder layers. We use the Cifar10[1] dataset for all our experiments.
|
|
Probabilistic creation and rendering of climbing plants
Rishubh Parihar,
Aniket Dashpute
Implemented a climbing heuristic to render climbers on required objects in a scene to enhance content creation
for computer graphics applications. Built a graph which is a minimal abstract representation of plant for production
of leaves and branches. Traversed the nodes in graph with geometry with materials and textures that can be rendered
|
|
Hierarchical Modelling of 3D animatable frog and key-frame animation
Rishubh Parihar,
Aniket Dashpute
Modeled a frog to be represented as a hierarchical model with an articulated structure
Created animation module by interpolating key frames with diffuse, specular and ambient components
Made an interactive game with multiple frogs which run behind use controlled insects
|
|
Recursive ray tracer for rendering simplistic scenes
Rishubh Parihar,
Aniket Dashpute
Implemented recursive ray tracing to generate an image of virtually generated 3D model by tracing path of light through pixels
Implemented global illumination model with reflection, refraction & shadows and local illumination model with diffuse, specular and ambient components
|
This website template is taken from here
|
|