top of page

DANSAEKHWA GENERATOR

IMAGE TRANSFORMATION TASK ON THE DANSAEKHWA DATASETS

Technological advances have brought tremendous changes not only to our general lives but also to the domain of art. As technology continues to progress, it has been functioning as a provider of new tools and applications for the arts and design disciplines. Among myriad applications, one of the trending topics is image transformation tasks, such as image-to-image translation, and style transfer.


Style transfer conventionally refers to the technique of reconstructing images in the style of other images. With the introduction and development of neural stylization methods, it is now possible to effectively, and artistically transfer input images into the images of any desired styles. The technique has allowed people who do not have any prior knowledge in art or coding to confidently produce artistic images that they desire to create. Various applications of style transfer are now available, ranging from the CycleGAN, a variant of GANs (Generative Adversarial Networks), to Prisma, a popular photo-editing app.

 

Closely related to neural style transfer, pix2pix is a classic example of image-to-image translation networks. These networks use a conditional generative adversarial network (cGAN) to learn a mapping from an input image to output image. The networks also learn a loss function to train the mapping, thus demonstrating a more effective approach to solving image-to-image translation tasks compared to other traditional models.


 

The images below are the examples of the image-to-image translation tasks available with pix2pix:

Capture.JPG

Retrieved from the paper Image-to-Image Translation with Conditional Adversarial Nets by Isola et al.)

THE PROJECT

The project was conducted as a part of the Dansaekhwa Project with the particular objectives of pondering the meaning of a creator in the artistic practice. The code used in the project were heavily borrowed from pix2pix-tensorflow (https://github.com/affinelayer/pix2pix-tensorflow). Among several options provided by the Dansaekhwa members, "Correspondence조응", the painting series by Lee Ufan, and "Untitled무제", a unique Umber Blue painting series, by Yun Hyong-keun are selected for their formal appropriacy, a perceivable similarity between each painting in the collections.

Examples of the target (original) images from the Untitled series by Yun Hyong-keun.

Examples of the target (original) images from the Correspondence series by Lee Ufan.

Each dataset consists of 500 pairs of the images, of 400 pairs used for training and the rest used for the test, which have a dimension of 512 × 512 pixels. 1,000 handmade equivalent input images for target images were created for these particular translation tasks. The images below are the examples of the pairs in each dataset.

target1.jpg
target2.jpg

The major issue when implementing the vanilla pix2pix was the pronounced checkerboard artifacts, one of the hallmarks of the artificially synthesized images. In order to alleviate the artifacts, the code has been added to the original model. The inserted code was referenced from Deconvolution and Checkerboard Artifacts, the article by Odena et al.

The pronounced checkerboard artifacts in the images generated by the original model. Images partially enlarged.

After inserting the code to resize convolution layers using bicubic resampling, the model stably generates more naturalistic images without exhibiting the pronounced checkerboard pattern of artifacts. The two collections of images below are the selection of output images comparing the results generated by the vanilla pix2pix and by the revised pix2pix networks.

Comparison of generated images with (bottom) and without (top) applying bicubic interpolation.

Comparison of generated images with (bottom) and without (top) applying bicubic interpolation.

CONCLUSION

As the above results of image-to-image translation tasks with pix2pix have revealed, the development of the image transformation technology has made it possible for people to create artistic images with minimal guidelines. Thus, the potential of artificial intelligence lies in its role as a collaborator, not there to replace human presence but to assist them to create something new, and something beautiful in a way that is not comparable with that of the past. The engagement of artificial intelligence in the domain of art has brought huge differences in our way of thinking, and in the perception of art and such change leads to the development of the new aesthetics in the digital world. Along with the emergence of altered perception, and of new methods of making art, human creativity will bloom elegantly with the help of artificial intelligence.

References

Zhu, J., Park, T., Isola, P., & Efros, A. A. (2017). Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks. 2017 IEEE International Conference on Computer Vision (ICCV). doi:10.1109/iccv.2017.244

 

Isola, P., Zhu, J., Zhou, T., & Efros, A. A. (2017). Image-to-Image Translation with Conditional Adversarial Networks. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). doi:10.1109/cvpr.2017.632

bottom of page