Skip to content

Latest commit

 

History

History
72 lines (43 loc) · 6.57 KB

README.md

File metadata and controls

72 lines (43 loc) · 6.57 KB

Collaborative Interactive Evolution of Art in the Latent Space of Deep Generative Models

This project was originally created as a master thesis at the VU Amsterdam and contains the implementation of the following paper:

Collaborative Interactive Evolution of Art in the Latent Space of Deep Generative Models
Ole Hall & Anil Yaman
https://doi.org/10.1007/978-3-031-56992-0_13

Abstract: Generative Adversarial Networks (GANs) have shown great success in generating high quality images and are thus used as one of the main approaches to generate art images. However, usually the image generation process involves sampling from the latent space of the learned art representations, allowing little control over the output. In this work, we first employ GANs that are trained to produce creative images using an architecture known as Creative Adversarial Networks (CANs), then, we employ an evolutionary approach to navigate within the latent space of the models to discover images. We use automatic aesthetic and collaborative interactive human evaluation metrics to assess the generated images. In the human interactive evaluation case, we propose a collaborative evaluation based on the assessments of several participants. Furthermore, we also experiment with an intelligent mutation operator that aims to improve the quality of the images through local search based on an aesthetic measure. We evaluate the effectiveness of this approach by comparing the results produced by the automatic and collaborative interactive evolution. The results show that the proposed approach can generate highly attractive art images when the evolution is guided by collaborative human feedback.

It contains a Creative Adversarial Network (CAN) for the generation of artworks. Resolutions of 64 x 64 and 256 x 144 (16:9) are available. The models are implemented with PyTorch. It further contains an evolutionary computing approach to control the latent variables from which the CAN's generator creates images. For this, the evolutionary algorithm framework DEAP was used. A local search following the work of Roziere et al. (2020) was implemented as mutation. To determine the quality of the art images, an automatic image evaluation metric based on Neural Image Assesment (NIMA) trained on Aesthetic Visual Analysis dataset (AVA) was applied. In addition, an interactive (collaborative) evaluation was implemented. The images were upsampled using The Laplacian Pyramid Super-Resolution Network (LapSRN).

Examples of the Collaborative Evolution

WikiArt Dataset

The publicly available WikiArts dataset was resized and used as training data. However, the models can also be trained on any other dataset. The wikiarts dataset can be downloaded from this repository

References

(1) Creative Adversarial Network

Creative Adversarial Network (CAN): Elgammal, A., Liu, B., Elhoseiny, M., Mazzone, M.: CAN: Creative adversarial networks, generating ”art” by learning about styles and deviating from style norms. arXiv:1706.07068 (2017)

The implementation of the CAN was based on this repository and also inspired by this one.

(2) Automatic and Collaborative Evolution

Neural Image Assesment (NIMA): Talebi, H., Milanfar, P.: NIMA: Neural image assessment. IEEE Transactions on Image Processing 27(8), 3998–4011 (2018)

Aesthetic Visual Analysis (AVA): Murray, N., Marchesotti, L., Perronnin, F.: AVA: A large-scale database for aesthetic visual analysis. In: IEEE Conference on Computer Vision and Pattern Recognition. pp. 2408–2415 (2012)

This implementation was used for NIMA without changes (the Inception ResNet v2 model, weights can be dowloaded here).

Laplacian Pyramid Super-Resolution Network (LapSRN): Lai, W.S., Huang, J.B., Ahuja, N., Yang, M.H.: Deep laplacian pyramid networks for fast and accurate super-resolution. In: IEEE Conferene on Computer Vision and Pattern Recognition (2017)

Upsampling was done using OpenCV, the pretrained model can be downloaded here.

Local Search: Roziere, B., Teytaud, F., Hosu, V., Lin, H., Rapin, J., Zameshina, M., Teytaud, O.: EvolGAN: Evolutionary generative adversarial networks. In: Proceedings of the Asian Conference on Computer Vision (2020)

Usage

To train the GAN, the use of GPU is recommended, especially for the higher resolution. As soon as a trained generator is provided, CPU is sufficient. The requirement.txt file contains all requirements (PyTorch is thereby only installed for CPU though).

Everything concerning CAN training can be found in the corresponding folder. In parameter.py, the overall parameters can bet set. train.py expects a folder containing a training dataset already in the appropriate size. The notebook offers an analysis of the training results.

For the evolutionary computing approach, there are several notebooks whose names are descriptive. All of them require a trained generator, most of them further require the availability of a pretrained NIMA model and a pretrained LapSRN model.

Citation

If you find this useful, please cite:

@InProceedings{10.1007/978-3-031-56992-0_13,
author="Hall, Ole and Yaman, Anil",
editor="Johnson, Colin and Rebelo, S{'e}rgio M. and Santos, Iria",
title="Collaborative Interactive Evolution of Art in the Latent Space of Deep Generative Models",
booktitle="Artificial Intelligence in Music, Sound, Art and Design",
year="2024",
publisher="Springer Nature Switzerland",
address="Cham",
pages="194--210",
doi="10.1007/978-3-031-56992-0_13"
}