Show simple item record

dc.contributor.authorWilliams, A
dc.contributor.authorLattner, S
dc.contributor.authorBarthet, M
dc.contributor.authorMusic Recommender Systems Workshop at the 17th ACM Conference on Recommender Systems
dc.contributor.editorFerraro, A
dc.contributor.editorKnees, P
dc.contributor.editorQuadrana, M
dc.contributor.editorYe, T
dc.contributor.editorGouyon, F
dc.date.accessioned2023-10-24T08:37:03Z
dc.date.available2023-08-28
dc.date.available2023-10-24T08:37:03Z
dc.date.issued2023-09-19
dc.identifier.urihttps://qmro.qmul.ac.uk/xmlui/handle/123456789/91539
dc.description.abstractWhile some artists are involved in both domains, the creation of music and artwork require different skill sets. The development of deep generative models for music and image generation has potential to democratise these mediums and make multi-modal creation more accessible for casual creators and other stakeholders. In this work, we propose a co-creative pipeline for the generation of images to accompany a musical piece. This pipeline utilises state-of-the-art models for music-to-text, image-to-text, and subsequently text-to-image generation to recommend, via generation, visuals for a piece of music that are informed not only by the audio of a musical piece, but also a user-recommended corpus of artworks and prompts to give a meaningful grounding in the generated material. We demonstrate the potential of our pipeline using a corpus of material from artists with strongly connected visual and musical identities, and make it available in the form of a Python notebook for users to easily generate their own musical and visual compositions using their chosen corpus - available here: https://github.com/alexjameswilliams/Music-Text-To-Image-Generationen_US
dc.publisherACMen_US
dc.subjectComputational Creativityen_US
dc.subjectGenerative AIen_US
dc.subjectImage Generationen_US
dc.subjectMusic Taggingen_US
dc.subjectPrompt Engineeringen_US
dc.subjectVisual Recommendationen_US
dc.titleSound-and-Image-informed Music Artwork Generation Using Text-to-Image Modelsen_US
dc.typeConference Proceedingen_US
pubs.author-urlhttps://orcid.org/0000-0003-2387-6876en_US
pubs.notesNot knownen_US
pubs.place-of-publicationNew York, NY, USAen_US
pubs.publication-statusPublished onlineen_US
dcterms.dateAccepted2023-08-28
qmul.funderUKRI Centre for Doctoral Training in Artificial Intelligence and Music::Engineering and Physical Sciences Research Councilen_US
qmul.funderUKRI Centre for Doctoral Training in Artificial Intelligence and Music::Engineering and Physical Sciences Research Councilen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record