Audio Metaphor(AUME)
Year: 2012-2024
Medium: Sound art, Interactive internet site, prompt-based interface.
Credits: Miles Thorogood, Philippe Pasquier.
Artistic direction, design, development: Miles Thorogood
Artistic direction, design: Philippe Pasquier
Development: Renaud Bougueng Tchemeube, Ge Liu, Jonas Kranabeter.
Description:
Audio Metaphor (AuMe) is a research project aimed at designing new methodologies and tools for sound design and composition practices in film, games, and sound art. Through this project, we have identified the processes involved in working with audio recordings in creative environments, addressing these in our research by implementing computational systems that can assist human operations.
We have successfully developed Audio Metaphor for the retrieval of audio file recommendations from natural language texts, and even used phrases generated automatically from Twitter to sonify the current state of Web 2.0. Another significant achievement of the project has been in the segmentation and classification of environmental audio with composition-specific categories, which were then applied in a generative system approach. This allows users to generate sound design simply by entering textual prompts.
As we direct Audio Metaphor further toward perception and cognition, we will continue to contribute to the music information retrieval field through environmental audio classification and segmentation. The project will continue to be instrumental in the design and implementation of new tools for sound designers and artists.
See more information on the website audiometaphor.ca.
Members
Miles Thorogood, Jianyu Fan, Philippe Pasquier, Arne Eigenfeldt, Renaud Bougueng, Ge Liu
Research papers and Posters
Thorogood, M. (2021) Developing a Sound Design Creative AI Methodology. Doing Research in Sound Design, 224-237. Focal Press.
Fan J., Yang Y-H., Dong K., Pasquer, P. (2020). A Comparative Study of Western and Chinese Classical Music based on Soundscape Models. International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Barcelona, Spain.
Fan J., Nichols E., Tompkins D., Méndez A. E. M., Elizalde B., Pasquer, P. (2020). Multi-label Sound Event Retrieval Using a Deep Learning-based Siamese Structure with a Pairwise Presence Matrix. International Conference on Acoustics, Speech, and Signal Processing, Barcelona, Spain.
Thorogood, M., Fan, J., Pasquer, P. (2019), A Framework for Computer-Assisted Sound Design Systems Supported by Modelling Affective and Perceptual Properties of Soundscapes. Journal of New Music Research.
Thorogood, M., & Pasquier, P. (2019) Soundscape Online Databases State of the Art and Challenges. In Foundations in Sound Design for Interactive Media: A Multidisciplinary Approach, 333. Routledge.
Thorogood, M. (2019) Soundscape Generation Systems. In Foundations in Sound Design for Interactive Media (pp. 259-273). Routledge.
Fan J., Thorogood, M., Tatar, K., Paquier, P. (2018). Quantitative Analysis of the Impact on Perceived Emotion of Soundscape Recordings. Sound and Music Computing (SMC)
Fan, J., Tung, F., Li, W., Pasquer, P. (2018). Soundscape Emotion Recognition via Deep Learning. Sound and Music Computing (SMC), 2018
Fan, J., Thorogood, M., and Pasquier, P. (2017). Emo-Soundscapes - A Dataset for Soundscape Emotion Recognition. Proceedings of the International Conference on Affective Computing and Intelligent Interaction.
Fan, J., Tatar, K., Thorogood, M., and Pasquier, P. (2017). Ranking-based Emotion Recognition for Experimental Music. Proceedings of the International Symposium on Music Information Retrieval, 2017.
Thorogood, Miles, Jianyu Fan and Philippe Pasquier. "Soundscape Audio Signal Classification and Segmentation Using Listeners Perception of Background and Foreground Sound". Journal of the Audio Engineering Society. Special Issue (Intelligent Audio Processing, Semantics, and Interaction), Oct 2016.
Fan, Jianyu, Miles Thorogood, and Philippe Pasquier. "Automatic Soundscape Affect Recognition Using A Dimensional Approach". Journal of the Audio Engineering Society. Special Issue (Intelligent Audio Processing, Semantics, and Interaction), Oct 2016.
Thorogood, M., Fan, J., Pasquier, P. BF-Classifier: Background/Foreground Classification and Segmentation of Soundscape Recordings. In Proceedings of the 10th Audio Mostly Conference, Greece, 2015.
Fan, J., Thorogood, M., Riecke, B., Pasquier, P. Automatic Recognition of Eventfulness and Pleasantness of Soundscape. In Proceedings of the 10th Audio Mostly Conference, Greece, 2015.
Eigenfeldt, A., Thorogood, M., Bizzocchi, J., Pasquier, P. MediaScape: Towards a Video, Music, and Sound Metacreation. Journal of Science and Technology of the Arts 6, 2014.
Thorogood, M, Pasquier, P., and Eigenfeldt, A. (2012). "Audio Metaphor: Audio Information Retrieval for Soundscape Composition" Sound and Music Computing (SMC). Copenhagen, Denmark.
Thorogood, M., Pasquier, P. (2013). "Computationally Generated Soundscapes with Audio Metaphor" In Proceedings of the 4th International Conference on Computational Creativity (ICCC). Sydney, Australia.
Thorogood, M., Pasquier, P. (2013). "Impress: A Machine Learning Approach to Soundscape Affect Classification for a Music Performance Environment" Proceedings of the 13th International Conference on New Interfaces for Musical Expression (NIME). Daejeon + Seoul, Korea Republic .
Bizzochi, Jim, Arne Eigenfeldt, Miles Thorogood and Justine Bizzochi. "Generating Affect: Applying Valence and Arousal values to a unified video, music, and sound generation system". Generative Art Conference. 2015. 308 - 318
Bizzochi, Jim, Arne Eigenfeldt, Philippe Pasquier and Miles Thorogood. "Seasons II: a case study in Ambient Video, Generative Art, and Audiovisual Experience". Electronic Literature Organization Conference. British Columbia, Canada. Jun, 2016. Electronic Literature Organization.