基于特征向量的音樂(lè)情感分析的研究
發(fā)布時(shí)間:2018-10-23 08:46
【摘要】:隨著當(dāng)今社會(huì)的迅速信息化,各種多媒體信息資料飛速發(fā)展。音樂(lè)作為一門藝術(shù),已經(jīng)成為人類生活中必備的部分。一直以來(lái),音樂(lè)都是人們表達(dá)情感的渠道,可以為歡樂(lè)而歌,可以為悲傷而唱。如今紙上的音樂(lè)已經(jīng)不能夠滿足音樂(lè)的保存、檢索以及音樂(lè)人之間的交流。隨著信息時(shí)代的到來(lái),計(jì)算機(jī)音樂(lè)的研究成了一個(gè)新的課題。讓計(jì)算機(jī)完成我們?nèi)祟惸軌蛲瓿傻氖虑橐恢笔侨藗冊(cè)噲D努力的方向。目前,我們可以通過(guò)計(jì)算機(jī)進(jìn)行音樂(lè)的播放、制作和存儲(chǔ)等,通過(guò)計(jì)算機(jī)對(duì)音樂(lè)的情感進(jìn)行分析也漸漸興起,使計(jì)算機(jī)能夠通過(guò)“聽(tīng)”音樂(lè)自動(dòng)識(shí)別出音樂(lè)所表達(dá)的情感。本文就音樂(lè)情感自動(dòng)分析做了深入的研究。本文的音樂(lè)情感分析模型由三個(gè)部分構(gòu)成:音樂(lè)特征向量模型、音樂(lè)情感模型和分類認(rèn)知模型。音樂(lè)特征向量模型是由從音樂(lè)中提取的一些特征組成的一個(gè)八維向量。在音樂(lè)特征向量模型的部分,本文在介紹了旋律面積的概念之后,定義了音樂(lè)能量的概念,并提出了自己的方法,即利用音樂(lè)能量為音樂(lè)劃分樂(lè)段,針對(duì)每個(gè)樂(lè)段使用數(shù)字音樂(lè)特征提取技術(shù)提取樂(lè)段的速度、旋律的方向、力度、節(jié)拍、節(jié)奏變化、大三度、小三度和音色等八個(gè)特征,然后利用音樂(lè)情感模型和分類認(rèn)知模型對(duì)每個(gè)樂(lè)段的情感進(jìn)行分析。音樂(lè)情感模型是音樂(lè)情感的描述,本文介紹了幾種研究者常用的音樂(lè)情感模型,包括Hevner情感環(huán)、Thayer情感模型和情感語(yǔ)義模型等等,并對(duì)這些模型的優(yōu)缺點(diǎn)進(jìn)行了比較。我們將Hevner情感環(huán)與情感語(yǔ)義模型相結(jié)合,得到了由Hevner情感環(huán)中的八大類情感描述所構(gòu)成的情感向量模型,并將該模型作為本文實(shí)驗(yàn)所用的情感模型。分類認(rèn)知模型是通過(guò)算法將音樂(lè)特征模型映射到情感模型,即分類認(rèn)知的過(guò)程是一個(gè)模式識(shí)別的過(guò)程。在分類認(rèn)知模型部分,簡(jiǎn)單介紹了幾種模式識(shí)別方法并對(duì)它們的優(yōu)缺點(diǎn)進(jìn)行對(duì)比之后,選用BP神經(jīng)網(wǎng)絡(luò)作為本文的認(rèn)知模型。針對(duì)音樂(lè)情感分析的需求,本文對(duì)BP神經(jīng)網(wǎng)絡(luò)的學(xué)習(xí)過(guò)程做了改進(jìn),使其能夠更加符合音樂(lè)情感分析的主觀性的特點(diǎn)。最后,本文將上述三部分自然地結(jié)合起來(lái),構(gòu)成了一個(gè)完整的音樂(lè)情感分析模型。之后,對(duì)該模型的功能和性能進(jìn)行了實(shí)驗(yàn)驗(yàn)證,并將實(shí)驗(yàn)結(jié)果與已有研究的實(shí)驗(yàn)成果進(jìn)行比較,結(jié)果顯示,使用本文所提出的方法構(gòu)建的音樂(lè)情感分析模型能夠較好地對(duì)數(shù)字音樂(lè)進(jìn)行情感分析,并且與已有成果相比,具有更高的準(zhǔn)確率。
[Abstract]:With the rapid development of information, various multimedia information materials are developing rapidly. As an art, music has become an essential part of human life. Music has always been a channel for people to express their feelings, to sing for joy and to sing for sorrow. Today, the music on paper can no longer satisfy the preservation, retrieval and communication between musicians. With the arrival of the information age, the research of computer music has become a new subject. Making computers do what we humans can do has always been the direction of our efforts. At present, we can play, make and store music by computer, and analyze the emotion of music by computer, so that the computer can automatically recognize the emotion expressed by music through listening to music. This article has done the thorough research to the music emotion automatic analysis. The music emotion analysis model consists of three parts: music feature vector model, music emotion model and classified cognitive model. The music feature vector model is an eight-dimensional vector composed of some features extracted from music. In the part of the music feature vector model, after introducing the concept of melodic area, this paper defines the concept of music energy, and puts forward its own method, that is, using music energy to divide music segments. For each segment, using digital music feature extraction technology to extract eight features, such as speed, direction of melody, intensity, rhythm, rhythm change, big third degree, small third degree and timbre, etc. Then the emotion of each segment is analyzed by using musical emotion model and classified cognitive model. Music emotion model is the description of music emotion. This paper introduces several musical emotion models commonly used by researchers, including Hevner emotional loop, Thayer emotional model and emotional semantic model, and compares the advantages and disadvantages of these models. We combine the Hevner emotional loop with the affective semantic model and obtain the emotional vector model which is composed of eight kinds of affective description in the Hevner emotional loop and use this model as the emotional model used in this experiment. Classifying cognitive model is to map music feature model to affective model through algorithm, that is, the process of classifying cognition is a process of pattern recognition. In the part of classified cognitive model, several pattern recognition methods are briefly introduced and their advantages and disadvantages are compared. Then BP neural network is selected as the cognitive model in this paper. According to the demand of music emotion analysis, this paper improves the learning process of BP neural network, so that it can more accord with the subjective characteristics of music emotion analysis. Finally, the above three parts are naturally combined to form a complete musical emotional analysis model. After that, the function and performance of the model are verified by experiments, and the experimental results are compared with the existing experimental results. The results show that, Using the method proposed in this paper, the music emotion analysis model can be used to analyze the digital music emotion better, and compared with the existing results, the model has a higher accuracy.
【學(xué)位授予單位】:西安電子科技大學(xué)
【學(xué)位級(jí)別】:碩士
【學(xué)位授予年份】:2014
【分類號(hào)】:TN912.3
本文編號(hào):2288731
[Abstract]:With the rapid development of information, various multimedia information materials are developing rapidly. As an art, music has become an essential part of human life. Music has always been a channel for people to express their feelings, to sing for joy and to sing for sorrow. Today, the music on paper can no longer satisfy the preservation, retrieval and communication between musicians. With the arrival of the information age, the research of computer music has become a new subject. Making computers do what we humans can do has always been the direction of our efforts. At present, we can play, make and store music by computer, and analyze the emotion of music by computer, so that the computer can automatically recognize the emotion expressed by music through listening to music. This article has done the thorough research to the music emotion automatic analysis. The music emotion analysis model consists of three parts: music feature vector model, music emotion model and classified cognitive model. The music feature vector model is an eight-dimensional vector composed of some features extracted from music. In the part of the music feature vector model, after introducing the concept of melodic area, this paper defines the concept of music energy, and puts forward its own method, that is, using music energy to divide music segments. For each segment, using digital music feature extraction technology to extract eight features, such as speed, direction of melody, intensity, rhythm, rhythm change, big third degree, small third degree and timbre, etc. Then the emotion of each segment is analyzed by using musical emotion model and classified cognitive model. Music emotion model is the description of music emotion. This paper introduces several musical emotion models commonly used by researchers, including Hevner emotional loop, Thayer emotional model and emotional semantic model, and compares the advantages and disadvantages of these models. We combine the Hevner emotional loop with the affective semantic model and obtain the emotional vector model which is composed of eight kinds of affective description in the Hevner emotional loop and use this model as the emotional model used in this experiment. Classifying cognitive model is to map music feature model to affective model through algorithm, that is, the process of classifying cognition is a process of pattern recognition. In the part of classified cognitive model, several pattern recognition methods are briefly introduced and their advantages and disadvantages are compared. Then BP neural network is selected as the cognitive model in this paper. According to the demand of music emotion analysis, this paper improves the learning process of BP neural network, so that it can more accord with the subjective characteristics of music emotion analysis. Finally, the above three parts are naturally combined to form a complete musical emotional analysis model. After that, the function and performance of the model are verified by experiments, and the experimental results are compared with the existing experimental results. The results show that, Using the method proposed in this paper, the music emotion analysis model can be used to analyze the digital music emotion better, and compared with the existing results, the model has a higher accuracy.
【學(xué)位授予單位】:西安電子科技大學(xué)
【學(xué)位級(jí)別】:碩士
【學(xué)位授予年份】:2014
【分類號(hào)】:TN912.3
【參考文獻(xiàn)】
相關(guān)碩士學(xué)位論文 前1條
1 鐘子岳;基于數(shù)據(jù)挖掘技術(shù)的音樂(lè)風(fēng)格分類方法的研究[D];南昌大學(xué);2013年
,本文編號(hào):2288731
本文鏈接:http://sikaile.net/kejilunwen/wltx/2288731.html
最近更新
教材專著