EvoMUSART Accepted Papers

Long talks

  • Prompt and Circumstances: Evaluating the Efficacy of Human Prompt Inference in AI-Generated Art
    Khoi Trinh, Scott Seidenberger, Joseph Spracklen, Raveen Wijewickrama, Bimal Viswanath, Murtuza Jadliwala and Anindya Maiti
  • Asɛmpayɛtsia: An Afrocentric Framework for Computational Creativity in Sound and Image
    Nana Amowee Dawson
  • Algorithms for Collaborative Harmonization
    Eyal Briman, Nimrod Talmon and Eyal Leizerovich
  • A Novel Diffusion Model based Approach for Sleep Music Generation
    Timo Hromadka, Kevin Monteiro and Sam Nallaperuma-Herzberg
  • Digital Artists’ Perceptions of Generative AI in South Asia: Insights from the South Asian Digital Arts Archive (SADA)
    Najam-Ul Assar and Megan Smith
  • A Dataset for Automatic Vocal Mode Classification
    Reem Hinrichs, Sonja Stephan, Alexander Lange and Jörn Ostermann
  • EvoLiveDJ: An LLM-Based Agentic System for Interactive Evolutionary Live Music Performance
    Kamer Ali Yuksel and Hassan Sawaf
  • AI Co-Artist: An LLM-Powered System for Interactive GLSL Shader Animation Evolution
    Kamer Ali Yuksel and Hassan Sawaf
  • Fluid Body: An Adaptive Embodied Sonification System for Cross-Cultural Performance
    Yuting Xue and Yueshen Wu
  • The Art That Poses Back: Assessing AI Pastiches after Contemporary Artworks
    Anca Daniela Dinu, Andreiana Mihail, Andra-Maria Florescu and Claudiu Dan Creangă
  • Generative musical exploration of astronomical catalogs
    Adrián García Riber
  • Artificial Intelligence in Music: Towards an Aesthetics of Co-Creation
    Anthony Grégoire, Serge Lacasse and Joanie Verviers-Deblois
  • Addressing Dataset Scarcity in Music Emotion Recognition with LLMs
    Janita Aamir and Patrick Donnelly

Short talks

  • Crystallizing Semantics: Mapping the Journey of Word Meaning in Language Models
    Himanshu Dwivedi
  • Quantum Latent Spaces for Symbolic Music Generation
    Sanjay Majumder and Neal Anderson
  • Multi-Objective Evolution of Diffusion Model Prompt Embeddings using CLIP-IQA
    Marcel Salvenmoser and Michael Affenzeller
  • Music In The Age Of Artificial Intelligence: Meaning And Creativity From A Complex-Systems Perspective
    Güncel Gürsel Artıktay
  • End-to-End Song Structure Segmentation via Encoder–Decoder Network Architecture and Hand-Crafted Features
    Phan Son, Nguyen Nghi and Lam Pham
  • Probing for Advanced Music Theory Concepts in Generative Music Models
    Derek Kwan and Patrick Donnelly
  • Generative Artificial Intelligence, Musical Heritage and the Construction of Peace Narratives: A Case Study in Mali
    Michael Leventhal, Nouhoum Coulibaly, Ousmane Ly and Ousmane Goro
  • Weather Sonification via a Latent Emotion Space: A Deep Learning Approach
    Takeshi Matsumura and Gavin Pringle
  • Life Beings: Living Quantum Art
    Alain Lioret, Kamal Hennou and Florentin Eraud
  • Mapping artificial neural networks’ processing data in audiovisual artworks
    Tanguy Pocquet, Ricardo Climent and Richard Allmendinger
  • EvoArtist: A Visual LLM-Driven Agentic AI Framework for Autonomous Design Evolution
    Kamer Ali Yuksel and Hassan Sawaf
  • Decoding Emotions: Multimodal integration of deep embeddings, lyrics and music-aware cues
    Alessia Novacco, Francesca Gasparini, Giulia Rizzi and Aurora Saibene
  • Classifying Audio Timbre Without Audio Using Text-only Training
    Peter McCabe and Patrick Donnelly
  • Segmentation-Free Sound Hybridization for Creative Outcomes
    Pasquale Mainolfi and Carmine Emanuele Cella
  • LoopMatcher: Proof-of-Concept for AI-Assisted Music Loop Search
    Subhrojyoti Roy Chaudhuri, Sai Pranav Madupu, Krishna Teja Guru Sai and Vikram Jamwal