VRYNT

GAN based NFT Builder/Generation for
NFT Marketplace

VRYNT (HashChi) wanted to create a GAN based NFT generation platform for allowing their community members to easily create/build unique NFTs by using components (latent space) and buy/sell in their Marketplace

Discuss Your Project

Scope of work

Document outlines specific project goals, tasks, timelines, and resources, ensuring a clear understanding of project boundaries. It acts as a roadmap, guiding stakeholders and facilitating successful outcomes
Artist Backend Collection Creation Interface
Collector NFT Builder Creation Tool
Gallery and Component trading marketplace
Smart Contracts to facilitate asset creation and some financial transactions

Problem 😨

Latency
Time
It was expected to reduce latency, for ref. StyleGAN2 takes approx 10 minutes for a 1024 pixel image for projection/rendering/inference on  but ADA drastically reduces this to 180 seconds (3 min approx.)

Research & Analysis

Rigorous research and data-driven insights drive success. Leverage our expertise, cutting-edge tools, and meticulous approach for informed choices in a dynamic landscape

Generative Modeling & GANs: Unleashing AI Creativity in Images

Generative Modeling is an unsupervised learning task that automates the discovery of patterns in input data, enabling the generation of new images or variations. The revolutionary Generative Adversarial Networks (GANs) model, introduced by Ian Goodfellow in 2014, excels at creating fake images resembling real data. With a dynamic interplay between its generator and discriminator sub-models, GANs push the boundaries of AI creativity, overcoming earlier limitations and achieving training stability, fidelity, and control over image features. DCGAN, ProGAN, and BigGAN are prominent GAN architectures that further enhance the generation of high-resolution, diverse images.

What we identify :

For our use case StyleGAN2 & StyleGAN-ADA is best fitted as it can generate high resolution images with diversity and have control over image features. The Stylegan2 model allows mapping from latent codes to images and vice versa. The Latter part is known as Image Inversion which allows any similar domain image w.r.t trained model to be embedded into latent space.

Ux Research Plans

Embedding new images into the Latent space

The first step for image manipulation in GANs is to be able to map a given image into the latent space. A popular approach to achieve this is to train an encoder to map the image into the latent space. After training the generator using GAN loss, we freeze the generator(decoder) weights and train an encoder that maps images to the latent space. The generator then generates a synthetic image based on the predicted latent. The encoder network is trained like an AutoEncoder by comparing the original image with the generator’s output. StyleGAN model architecture has built in Autoencoder to invert images in latent space.

Feature entanglement in GANs

Stylegan introduced a mapping network to untangle image features, the goal of the mapping network is to convert the input latent vector into the intermediate vector whose different elements control different visual features. StyleGAN warps a space that can be sampled with a uniform or normal distribution into the latent feature space with disentangled feature space.The mapping function is implemented using 8-layer MLP (8-fully connected layers).The output of mapping network (w) then passes through a learned affine transformation (A) before passing into the synthesis network which uses the AdaIN (Adaptive Instance Normalization) module. This model converts the encoded mapping into the generated image.In order to have more control on the styles of the generated image, the synthesis network provides control over the style to different levels of details (or resolution) during image synthesis & upscaling from constans vector to resolution 1024 as shown in figure.These different levels are defined as

  • Coarse : resolution of  ( 4×4 – 8×8) – affects pose, general hair style, face shape, etc
  • Middle : resolution of (16×16 – 32×32) – affects finer facial features, hair style, eyes open/closed, etc
  • Fine : resolution of (64×64 – 1024×1024) – affects colors for (eye, hair and skin) and micro features.
Image Feature Manipulation

Image features control can be achieved by identifying latent direction using Principal component analysis (PCA) in activation space and manual labeling of each identified latent direction based upon changes appearing in the image

Style mixing is another way to manipulate image features by imparting source image features of target image. This involves combining two image ( source and target) latent vectors, more layers are combined, more source image look-alike target and vice versa.

Model Training

Training a new model using StyleGAN-ADA requires less images compared to stylegan which requires large no. of images as it uses image augmentation technique to increase dataset size and takes less time to train a new model with limited images. Training a stylegan ada model is a gpu intensive task , Frechet Inception Distance (FID) score is used to keep a check on how well a model is performing while training a new model from scratch.

Solution

VRYNT and Rapid Innovation joined hands to build an AI-based NFT creation space where platform users have creative freedom to manipulate image features, generating completely unique images using GANs based image modeling, and sell those created art as NFT. Artists can use the gene component provided with each domain image to manipulate image features, for example in landscape image domain , artists can start with selecting in-domain image of model such as bare landscape image and by using the gene component of landscape model new features can be added to image such as trees, pond/river, mountains etc. same can be seen in attached figure below. Artist have liberty to combine as many gene from component pack to create a unique art that can be listed as NFT in VRYNT marketplace. Also allows artists to get a transition image generated during the feature manipulation process recorded as GIF
Platform also allows users to upload images in certain image domains such as Human, manipulate features of input image. GANs based image models require large-scale image dataset & high compute resources to train a new model, Here we used available open-source models & fine-tuned them if the model lacked diversity & trained some model from scratch according to client needs. All image domain & genes components available on platform are listed below.
Image Domain
Landscape Oil Painting
Visionary Art
Modern art
Metal Album Art
Disney Cartoon Faces
New Abstract Art
Archillect
Night Sky
Texture
Scifi
Ponies
Painting MetFace
New Art
WikiArt
Human Face
Anime
Fursona
Landscapes Ada
Abstract art
Image Domain
50
51
55
51
52
51
51
51
57
51
59
55
51
50
57
93
54
51
51
Image Domain
1024
1024
1024
1024
1024
1024
1024
1024
1024
1024
1024
1024
1024
1024
1024
512
512
256
256

Low Fidelity Wireframes

Low Fidelity Wireframes are valuable tools that aid projects by facilitating rapid iteration, gathering early feedback, and emphasizing core functionalities. Cost-effective and easy to collaborate with, wireframes visualize project structures and reduce misinterpretation, ensuring a user-centric approach and a strong foundation for successful outcomes.

Architecture Design [ Production ]

Crafting a comprehensive blueprint, we build the foundation for scalable solutions, ensuring performance, security, and efficiency. Expertly chosen technologies and foresight for growth drive successful project realization

Designs

Finally, We have the high-fidelity screens of the application. Introducing resume, your one-stop solution to returning back to work successfully
Home
Marketplace
Home
Marketplace
📌 AI-ML Backend Tech Stack
  • Tensorflow
  • Pytorch
  • Celery
  • RabbitMQ
  • Python
  • AWS Services (S3, Cloudwatch, EC2)
  • Flask
  • CUDA Toolkit
  • Nginx
  • Docker
  • Sklearn
  • OpenCV
🎓Conclusion

VRYNT project has been completed and deployed on client server with remarkable 20+ GAN Models and 2 Million images in Gallery Collections for creating & blending unique and one of its kind NFTs to place in the marketplace.

});