This is an artificial intelligence project, with the objective of translating pieces of visual art into music, enabling any person to feel art regardless of their visual impairment.
For this purpose, we have developed a convolutional neural network that extracts the sentiment from a painting in terms of valence-arousal. Then, a MIDI audio file is picked from a dataset labelled by emotion, which combines classical music and songs from videogames. A MusicTransformer produces an entirely new song inspired on the selected track.
This package features a Streamlit application that brings all these pieces together.
Clone the repository and navigate into the folder:
git clone https://github.com/GANs-N-Roses/deploy.git art2music
cd art2music
This repository does not include a soundfount for filesize reasons. Download a soundfont and place it (a default.sf2 file) in the ./visualization/soundfonts/
directory. This soundfount will be used to render the output song.
Build and run the Docker containers:
docker-compose up
Visit the Streamlit web application (at http://localhost:8000), input an image and wait a few minutes to get a completely original, emotion-conditioned song generated
Pull requests are welcome. For major changes, please open an issue first to discuss what you would like to change.
Please make sure to update tests as appropriate.
Music Transformer: Generating Music with Long-Term Structure Huang, Cheng-Zhi Anna and Vaswani, Ashish and Uszkoreit, Jakob and Shazeer, Noam and Hawthorne, Curtis and Dai, Andrew M and Hoffman, Matthew D and Eck, Douglas, arXiv preprint arXiv:1809.04281, 2018
Studying emotion induced by music through a crowdsourcing game A. Aljanaki, F. Wiering, R. C. Veltkamp. Information Processing & Management, 2015.
Learning to Generate Music with Sentiment Ferreira, Lucas N. and Whitehead, Jim Proceedings of the Conference of the International Society for Music Information Retrieval ISMIR'19
WikiArt Emotions: An Annotated Dataset of Emotions Evoked by Art Saif M. Mohammad and Svetlana Kiritchenko In Proceedings of the 11th Edition of the Language Resources and Evaluation Conference (LREC-2018), May 2018, Miyazaki, Japan.