Javascript must be enabled to continue!
AI-Driven Music Generation and Emotion Conversion
View through CrossRef
With the integration of Generalized Adversarial Networks (GANs), Artificial Intelligence Generated Content (AIGC) overcomes algorithmic limitations, significantly enhancing generation quality and diversifying generation types. This advancement profoundly impacts AI music generation, fostering emotionally warm compositions capable of forging empathetic connections with audiences. AI interprets input prompts to generate music imbued with semantic emotions. This study aims to assess the accuracy of AI music generation in conveying semantic emotions, and its impact on empathetic audience connections. ninety audios were generated across three music-generated software (Google musicLM, Stable Audio, and MusicGen), using four emotion prompts (Energetic, Distressed, Sluggish, and Peaceful) based on the Dimensional Emotion Model, and two generated forms (text-to-music and music-to-music). Emotional judgment experiment involving 26 subjects were conducted, comparing their valance and arousal judgments of the audios. Through Multi-way variance analysis, the AI-music-generated software had a significant main effect on the accuracy of conversion. Due to the diversity of generated forms of MusicGen, it has a lower accuracy of conversion compared to Google musicLM and Stable Audio. There was a significant interaction effect of generated forms and emotion prompts on the accuracy of conversion. The differences in accuracy between emotion prompts in the form of text-to-music were statistically significant, except for the differences between the accuracy of Distressed and Peaceful. Compared with the generated form of text-to-music, the form of music-to-music showed statistically significant emotional conversion ability for low arousal. The diversity of AI software input elements (i.e., text or music) may affect the effectiveness of emotional expression in music generation. The ability of different software to convey different emotions according to different prompts was unsteady in the form of text-to-music. This study advance computer music co-composition and improvisation abilities, facilitating AI music applications in fields such as medical rehabilitation, education, psychological healing, and virtual reality experiences.
AHFE International
Title: AI-Driven Music Generation and Emotion Conversion
Description:
With the integration of Generalized Adversarial Networks (GANs), Artificial Intelligence Generated Content (AIGC) overcomes algorithmic limitations, significantly enhancing generation quality and diversifying generation types.
This advancement profoundly impacts AI music generation, fostering emotionally warm compositions capable of forging empathetic connections with audiences.
AI interprets input prompts to generate music imbued with semantic emotions.
This study aims to assess the accuracy of AI music generation in conveying semantic emotions, and its impact on empathetic audience connections.
ninety audios were generated across three music-generated software (Google musicLM, Stable Audio, and MusicGen), using four emotion prompts (Energetic, Distressed, Sluggish, and Peaceful) based on the Dimensional Emotion Model, and two generated forms (text-to-music and music-to-music).
Emotional judgment experiment involving 26 subjects were conducted, comparing their valance and arousal judgments of the audios.
Through Multi-way variance analysis, the AI-music-generated software had a significant main effect on the accuracy of conversion.
Due to the diversity of generated forms of MusicGen, it has a lower accuracy of conversion compared to Google musicLM and Stable Audio.
There was a significant interaction effect of generated forms and emotion prompts on the accuracy of conversion.
The differences in accuracy between emotion prompts in the form of text-to-music were statistically significant, except for the differences between the accuracy of Distressed and Peaceful.
Compared with the generated form of text-to-music, the form of music-to-music showed statistically significant emotional conversion ability for low arousal.
The diversity of AI software input elements (i.
e.
, text or music) may affect the effectiveness of emotional expression in music generation.
The ability of different software to convey different emotions according to different prompts was unsteady in the form of text-to-music.
This study advance computer music co-composition and improvisation abilities, facilitating AI music applications in fields such as medical rehabilitation, education, psychological healing, and virtual reality experiences.
Related Results
Owner Bound Music: A study of popular sheet music selling and music making in the New Zealand home 1840-1940
Owner Bound Music: A study of popular sheet music selling and music making in the New Zealand home 1840-1940
<p>From 1840, when New Zealand became part of the British Empire, until 1940 when the nation celebrated its Centennial, the piano was the most dominant instrument in domestic...
What about males? Exploring sex differences in the relationship between emotion difficulties and eating disorders
What about males? Exploring sex differences in the relationship between emotion difficulties and eating disorders
Abstract
Objective: While eating disorders (ED) are more commonly diagnosed in females, there is growing awareness that men also experience ED and may do so in a different ...
Introduction: Autonomic Psychophysiology
Introduction: Autonomic Psychophysiology
Abstract
The autonomic psychophysiology of emotion has a long thought tradition in philosophy but a short empirical tradition in psychological research. Yet the past...
Advancing knowledge in music therapy
Advancing knowledge in music therapy
It is now over 20 years since Ernest Boyer – an educator from the US and, amongst other posts, President of the Carnegie Foundation for the Advancement of Teaching – published his ...
Music Video
Music Video
Music video emerged as the object of academic writing shortly after the introduction in the United States of MTV (Music Television) in 1981. From the beginning, music video was cla...
Welcome to the Robbiedome
Welcome to the Robbiedome
One of the greatest joys in watching Foxtel is to see all the crazy people who run talk shows. Judgement, ridicule and generalisations slip from their tongues like overcooked lamb ...
If I Had Possession over Judgment Day: Augmenting Robert Johnson
If I Had Possession over Judgment Day: Augmenting Robert Johnson
augmentvb [ɔːgˈmɛnt]1. to make or become greater in number, amount, strength, etc.; increase2. Music: to increase (a major or perfect interval) by a semitone (Collins English Dicti...
Dragutin Gostuški’s Television Narrative
Dragutin Gostuški’s Television Narrative
The selection of music combined with the text about music is very important for the effect on the viewer of the television music programs. The interaction between music and text tu...

