NVIDIA GauGAN Perhaps the Coolest GTC 2019 Demo

3
NVIDIA GauGAN Scene 1
NVIDIA GauGAN Scene 1

At GTC 2019, there were a number of great demos by both NVIDIA and its partners. One NVIDIA specifically briefed the press on doing a live demonstration before the Keynote session at San Jose State. NVIDIA GauGAN is a model trained on a myriad of photos. Its application and output are quite amazing. Given a rough sketch of what a composer wants a scene to include, it is able to generate a scene that incorporates those elements.

NVIDIA GauGAN at GTC 2019

The NVIDIA GauGAN demo at GTC 2019 was immensely interesting. My friends and family all know that I have, what can be termed as, “limited artistic skill.” More succinctly, I failed coloring in kindergarten and never recovered. The concept of creating a photorealistic scene with my drawing seemed farfetched to me before seeing this demo.

Here is the NVIDIA demo of GauGAN to give you an idea of what it can do:

By composing a scene based on what you want to see in different locations of the frame, GauGAN is able to fill in that scene. What is amazing is not just the basic nature of this capability. One can change the settings to different seasons or time of day and the network changes the image output. For example, if you have a tree GauGAN will give it leaves in the summer and just branches in the winter when the scene is covered in snow.

Here is an NVIDIA sample input and output.

NVIDIA GauGAN Scene 2
NVIDIA GauGAN Scene 2

I asked if I could try it myself to see if my poor artistic skills can be rectified and I was told no. Still, we saw a live demo before the GTC keynote this week. The sample images may be the best examples, but they capture the essence of what GauGAN does fairly well based on what we saw in the live demo.

Some of the finer details were a bit more spectacular. One example the company showed is that by adding a mountain and a lake element to the scene, GauGAN actually adds the mountain’s reflection to the lake’s surface. Changes like adding the lake, with reflection, happened single digit seconds after they were made, not hours after the fact. Frankly, a lot of this happened faster than standard Adobe Photoshop filters get applied on a high-end workstation.

There were still a few rough edges. Some of the boundaries did not look perfect using wide brush strokes. We were told there are still a few improvements but using a finer brush would have yielded better results, albeit not created at a pace fast enough for a live demo in front of an audience.

Learning More

You can learn more about GauGAN at the NVIDIA blog. I hope we get to actually use this at some point and it does not stay relegated to a research project.

3 COMMENTS

LEAVE A REPLY

Please enter your comment!
Please enter your name here

This site uses Akismet to reduce spam. Learn how your comment data is processed.