deepsing translates music into unique, machine generated images keeping the sentiment and the arroused emotions - create visual stories using audio as input
deepsing was born to materialize our idea of translating audio to images inspired by Futurama Holophoner. It works by performing attributed-based music-to-image translation and synthesizes visual stories according to the sentiment expressed by songs. The sentiment-aware generated images aim to induce the same feelings to the viewers, as the original song does, reinforcing the primary aim of music, i.e., communicating feelings.
Classify music segments based on valence and arousal
Map audio and sentiment features to image categories
Enchance images' sentiment using neural style transfer
Deliver unique visual stories using Generative Adversarial Networks