Responsive image
博碩士論文 etd-0703109-195627 詳細資訊
Title page for etd-0703109-195627
論文名稱
Title
一個於音樂資料庫中有效率的JMC節奏查詢之方法
An Efficient JMC Algorithm for the Rhythm Query in Music Databases
系所名稱
Department
畢業學年期
Year, semester
語文別
Language
學位類別
Degree
頁數
Number of pages
88
研究生
Author
指導教授
Advisor
召集委員
Convenor
口試委員
Advisory Committee
口試日期
Date of Exam
2009-06-12
繳交日期
Date of Submission
2009-07-03
關鍵字
Keywords
音樂資料庫、節奏、快速-緩慢、音樂時間序列
quick-slow, Music Databases, rhythm, music duration sequence
統計
Statistics
本論文已被瀏覽 5719 次,被下載 0
The thesis/dissertation has been browsed 5719 times, has been downloaded 0 times.
中文摘要
中文摘要
近年來,由於科技的進步,音樂的取得越來越容易,音樂也變得越來越普遍化。在我們周圍,各種各樣的音樂變得更加複雜及大量。音樂的爆炸性增加,迫使我們需要新的技術和工具,可以智慧地且自動地將音樂轉換成有用的資訊,並且將音樂分類(classification)到正確的類別。節奏(rhythm)查詢是音樂類別分類的基本技術,這也是重要的多媒體應用。最近,Christodoulakis 等人提出針對音樂時間序列(music duration sequence)以節奏來分類的CIRS 演算法。在CIRS 運算法的觀念中,一段節奏可視為一串由快速(Q)和緩慢(S)的字元組成序列,其涵意為音樂中音符的時間序列。在比對過程中,兩個連續的快速(Q)字元所組成的序列,可以組合成一個緩慢(S)字元,但是一個緩慢(S)字元不能分解成兩個連續的快速(Q)字元,這點也造成了演算法設計的困難。為了要以節奏來將音樂分類,CIRS 演算法找出MaxCover—在音樂時間序列中,被查詢之節奏所連續覆蓋(重疊地或相連地)的最長子序列。在CIRS 演算法中,可分為4 個主要步驟,以重複執行步驟2、3、4 的方式,在音樂時間序列中,針對每個不同的值掃描輸入的時間序列,找出所有的區域性MaxCover。最後以區域性MaxCover,計算出全域性MaxCover,這也是演算法最後的結果。但是,我們觀察到,CIRS 演算法,在步驟2、3、4 中產生出非必要的步驟及結果,由此會浪費演算法的執行時間。為了避免產生以上情形,我們提出JMC(Jumping by MaxCover)演算法,可以解決與CIRS 演算法相同的問題,漸增地產生MaxCover,並且提供一個刪除方法來降低執行花費。事實上,我們是從以相異時間值(different duration vale)X 所找出來的MaxCover MX,以及以相異時間值X 所截斷(cut)的時間序列,利用它們之間的關係,來減少以其他相異時間值Y 來執行演算法的花費,其Y < X。善用這項特性來降低執行時間,我們提出一個cut-sequence 的結構,並且漸增地更新它來計算出最後的全域MaxCover。透過上述方法,我們可以跳過許多步驟並且找出與CIRS 演算法相同的結果。最後,經由我們模擬結果,我們顯示出JMC 演算法在執行時間上比CIRS 演算法更快。而且當最大的相異時間值是均勻地分布在時間序列上,其執行時間會大幅縮減,這也是我們所提出JMC 演算法的最佳情形。
Abstract
In recent years, the music has become more popular due to the evolution of the technology. Various kinds of music around us become more complexity and huge. This explosive growth in the music has generated the urgent need for new techniques and tools that can intelligently and automatically transform the music into useful information, and classify the music into correct music groups precisely. The rhythm query is the fundamental technique in music genre classification and content-based retrieval, which are crucial to multimedia applications. Recently, Christodoulakis et al. has proposed the CIRS algorithm that can be used to classify music duration sequences according to rhythms. In the CIRS algorithm, a rhythm is represented by a sequence of “Quick” (Q) and “Slow” (S) symbols, which corresponds to the (relative) duration of notes, such that S = 2Q. In order to classify music by rhythms, the CIRS algorithm locates the MaxCover which is the maximum-length substring of the music duration sequence, which can be covered (overlapping or consecutively) by the rhythm query continuously. During the matching step, one S symbol in the rhythm query can be regarded as two consecutive Q symbols in the duration sequence, but the two consecutive Q symbols in the rhythm query can not be combined as one S symbol in the duration sequence. This definition causes the difficulty for designing the algorithm. The CIRS algorithm contains four steps and repeat Steps 2, 3, and 4 to get local MaxCover for each different duration value of the music duration sequence. Finally, the global MaxCover is computed. We observe that it will generate unnecessary results repeatedly among Steps 2, 3, and 4. Therefore, in this thesis, to avoid repeatedly processing Steps 2, 3, and 4 for each different duration value, we propose the JMC (Jumping-by-MaxCover) algorithm which provides a pruning strategy to find the MaxCover incrementally, resulting in the reducing of the processing cost. In fact, we can make use of the relationship between the MaxCover MX founded by a different duration value X, and use the duration sequences cut by such a different duration value X to reduce the unnecessary process for the other different duration value Y , where Y < X. To make use of this property to reduce the processing time, we propose a cut-sequence structure and update it incrementally to compute the final global MaxCover. In this way, we can skip many steps and find the same answer of the CIRS algorithm. From our simulation results, we show that the running time of the JMC algorithm could be shorter than that of the CIRS algorithm. When the largest different duration value is uniformly distributed in the duration sequence, the running time can be reduced hugely, which is the best case of our proposed JMC algorithm.
目次 Table of Contents
TABLE OF CONTENTS
Page
ABSTRACT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . i
LIST OF FIGURES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iv
LIST OF TABLES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii
1. Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1 Basic Music Elements . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1.1 Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1.2 The Pitch . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.1.3 Temporal Features . . . . . . . . . . . . . . . . . . . . . . . . 4
1.2 Music Information Retrieval . . . . . . . . . . . . . . . . . . . . . . . 6
1.2.1 Applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.3 Related Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
1.3.1 Pitch Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
1.3.2 Rhythm Analysis . . . . . . . . . . . . . . . . . . . . . . . . . 9
1.3.3 Query by Humming . . . . . . . . . . . . . . . . . . . . . . . . 9
1.3.4 The CIRS Algorithm . . . . . . . . . . . . . . . . . . . . . . . 12
1.4 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
2. A Survey of Music Information Retrieval . . . . . . . . . . . . . . . 19
2.1 The CIRS Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
2.2 L-tree . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
2.3 The Dynamic Time Warping . . . . . . . . . . . . . . . . . . . . . . . 26
2.4 The Hidden Markov Model . . . . . . . . . . . . . . . . . . . . . . . . 28
3. The Jumping-By-MaxCover Algorithm . . . . . . . . . . . . . . . . 32
3.1 Definitions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
3.1.1 Duration Sequences . . . . . . . . . . . . . . . . . . . . . . . . 32
3.1.2 The Rhythm Representation . . . . . . . . . . . . . . . . . . . 33
3.1.3 q-Match . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34
3.1.4 q-Cover . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36
3.1.5 The Maximal Coverability Problem . . . . . . . . . . . . . . . 37
3.2 The Proposed JMC Algorithm . . . . . . . . . . . . . . . . . . . . . . 38
3.3 Step 1: Finding All Occurrence of S . . . . . . . . . . . . . . . . . . . 39
3.4 Step 2: Transformation . . . . . . . . . . . . . . . . . . . . . . . . . . 44
3.5 Step 3: Finding Matchings . . . . . . . . . . . . . . . . . . . . . . . . 44
3.6 Step 4: Finding MaxCover . . . . . . . . . . . . . . . . . . . . . . . . 49
3.7 Step 5: Updating Cut-Sequence . . . . . . . . . . . . . . . . . . . . . 50
3.8 A Comparison . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
4. Performance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
4.1 Generation of Synthetic Data . . . . . . . . . . . . . . . . . . . . . . 57
4.2 Simulation Results of Synthetic Data . . . . . . . . . . . . . . . . . . 60
4.3 Performance Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . 68
5. Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
5.1 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
5.2 Future Works . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
BIBLIOGRAPHY . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72
參考文獻 References
[1] A. L. P. Chen, C. S. Iliopoulos, S. Michalakopoulos, and M. S. Rahman, “Implementation of Algorithms to Classify Musical Texts According to Rhythms,” Proc. 4th of Int. Sound and Music Computing Conf., pp. 11–13, 2007.

[2] H. C. Chen, Y. H. Wu, Y. C. Soo, and A. L. P. Chen, “Continuous Query Processing over Music Streams Based on Approximate Matching Mechanisms,” Multimedia Systems, Vol. 14, No. 1, pp. 51–70, June 2008.

[3] J. C. C. Chen and A. L. P. Chen, “Query by Rhythm: An Approach for Song Retrieval in Music Databases,” Proc. IEEE 8th Int. Workshop on Research Issues in Data Eng. : Continuous-Media Databases and Applications, pp. 139–146, 1998.

[4] L. Chen and B. G. Hu, “An Implementation of Web Based Query by Humming System,” IEEE Int. Conf. on Multimedia and Expo., pp. 1467–1470, 2007.

[5] M. Christodoulakis, C. S. Iliopoulos, and M. S. Rahman, “Identifying Rhythms in Musical Texts,” Int. Journal of Foundations of Computer Science, Vol. 19, No. 1, pp. 37–51, Feb. 2008.

[6] R. B. Dannenberg, W. P. Birmingham, G. P. Tzanetakis, C. P. Meek, N. P. Hu, and B. P. Pardo, “The MUSART Testbed for Query-by-Humming Evaluation,” Computer Music Journal, Vol. 28, No. 2, pp. 34–48, March 2004.

[7] S. Dixon, “On the Analysis of Musical Expression in Audio Signals,” Proc. of the SPIE Conf. on Storage and Retrieval for Media Database, pp. 122–132, 2003.

[8] J. S. Downie, “Music Information Retrieval,” Annual Review of Information Science and Technology, Vol. 37, pp. 295–340, 2003.

[9] A. Duda, A. Nurnberger, and S. Stober, “Towards Query by Singing/Humming on Audio Databases,” Proc. of the 8th Int. Conf. on Music Information Retrieval, pp. 331–334, 2007.

[10] A. Ghias, J. Logan, D. Chamberlin, and B. C. Smith, “Query by Humming-musical Information Retrieval in an Audio Database,” ACM Multimedia, pp. 231–236, 1995.

[11] C. L. Hsu, H. R. Lee, and J. S. R. Jang, “Continuous HMM and Its Enhancement for Singing/Humming Query Retrieval,” Proc. of the 6th Int. Conf. on Music Information Retrieval, pp. 1064–1069, 2005.

[12] J.-S. R. Jang and H.-R. Lee, “Hierarchical Filtering Method for Content-based Music Retrieval via Acoustic Input,” Proc. of the 9th ACM Int. Conf. on Multi- media, pp. 401–410, 2001.

[13] D. Little, D. Raffensperger, and B. Pardo, “A Query by Humming System That Learns from Experiences,” Proc. of the 8th Int. Conf. on Music Information Retrieval, pp. 23–27, 2007.

[14] B. Liu and Y. Li, “Linear Hidden Markov Model for Music Information Retrieval Based on Humming,” Proc. of IEEE Int. Conf. on Acoustics, Speech, and Signal Processing, Vol. 5, pp. 533–536, 2003.

[15] N. Liu, Y. Wu, and A. L. P. Chen, “Efficient kNN Search in Polyphonic Music Databases Using a Lower Bounding Mechanism,” ACM Multimedia Systems, Vol. 10, No. 6, pp. 513–528, Oct. 2005.

[16] J.-J. Nattiez, Music and Discourse: Toward a Semiology of Music. 1990.

[17] N. Orio, “Music Retrieval: A Tutorial and Review,” Foundations and Trends in Information Retrieval, Vol. 1, No. 1, pp. 1–96, Jan. 2006.

[18] D. M. Randel, The New Harvard Dictionary of Music. 1986.

[19] J. Shifrin, B. Pardo, A. Arbor, A. Arbor, and Birmingham, “HMM-based Musical Query Retrieval,” Proc. of the 2nd ACM/IEEE-CS Joint Conf. on Digital libraries, pp. 295–300, 2002.

[20] H.-H. Shih, S. S. Narayanan, and C.-C. J. Kuo, “An HMM-based Approach to Humming Transcription,” Proc. of IEEE Int. Conf. on Multimedia and Expo, Vol. 1, pp. 337–340, Aug. 2002.

[21] D. Tao, H. Liu, and X. Tang, “K-Box: A Query-by-Singing BasedMusic Retrieval System,” Proc. of ACM Int. Conf. on Multimedia, pp. 464–467, 2004.

[22] The Annual Int. Conf. on Music Information Retrieval, http://www.ismir.net/.

[23] E. Unal, E. Chew, P. G. Georgiou, and S. S. Narayanan, “Challenging Uncertainty in Query by Humming Systems: A Fingerprinting Approach,” IEEE Trans. on Audio, Speech, and Language Processing, Vol. 16, No. 2, pp. 359–371, Feb. 2008.

[24] Y. Zhu and D. Shasha, “Warping Indexes With Envelope Transforms for Query by Humming,” Proc. of the 2003 ACM SIGMOD Int. Conf. on Management of Data, pp. 181–192, 2003.
電子全文 Fulltext
本電子全文僅授權使用者為學術研究之目的,進行個人非營利性質之檢索、閱讀、列印。請遵守中華民國著作權法之相關規定,切勿任意重製、散佈、改作、轉貼、播送,以免觸法。
論文使用權限 Thesis access permission:校內校外均不公開 not available
開放時間 Available:
校內 Campus:永不公開 not available
校外 Off-campus:永不公開 not available

您的 IP(校外) 位址是 3.141.31.240
論文開放下載的時間是 校外不公開

Your IP address is 3.141.31.240
This thesis will be available to you on Indicate off-campus access is not available.

紙本論文 Printed copies
紙本論文的公開資訊在102學年度以後相對較為完整。如果需要查詢101學年度以前的紙本論文公開資訊,請聯繫圖資處紙本論文服務櫃台。如有不便之處敬請見諒。
開放時間 available 已公開 available

QR Code