Responsive image
博碩士論文 etd-0809107-170340 詳細資訊
Title page for etd-0809107-170340
論文名稱
Title
應用判別函式結合單類別支援向量機於多類別分類
Applying Discriminant Functions with One-Class SVMs for Multi-Class Classification
系所名稱
Department
畢業學年期
Year, semester
語文別
Language
學位類別
Degree
頁數
Number of pages
74
研究生
Author
指導教授
Advisor
召集委員
Convenor
口試委員
Advisory Committee
口試日期
Date of Exam
2007-07-26
繳交日期
Date of Submission
2007-08-09
關鍵字
Keywords
支援向量群聚、判別函式、單類別支援向量機、多類別資料分類
AdaBoost.M1, multi-class classification, One-class SVM, Discriminant function, Support vector clustering
統計
Statistics
本論文已被瀏覽 5798 次,被下載 0
The thesis/dissertation has been browsed 5798 times, has been downloaded 0 times.
中文摘要
在多類別分類的問題中,AdaBoost.M1演算法扮演了相當成功的角色,並在準確度方面有著不錯的提升效果。然而,此演算法假定所有基礎分類器於準確度方面都必須大於1/2,這在多類別分類的問題中是難以達到的。因此,我們提出一個新的演算法- AdaBoost.MK,它只要求在多類別分類問題中,每個分類器的正確率需比隨機亂猜的機率(1/k)還要高即可。
早期支援向量機在處理多類別資料分類問題時,都將它分解成數個兩類別的問題再加以解決,如此一來將需要花費大量的時間以及空間計算。為了減少時間與空間上的花費,我們提出了一個結合數個單類別支援向量機(one-class svm)以及判別函式(discriminant function)的方法來解決多類別資料分類的問題。
在本論文中,我們利用整合數個單類別支援向量機以及判別函式的方法並與AdaBoost.MK做結合來解決多類別資料分類的問題。實驗中,我們將使用來自UCI 和Statlog的真實資料,並與許多多類別分類的演算法做比較,包括利用支援向量群聚(support vector clustering)的方法與AdaBoost.M1結合單類別支援向量機的方法。
Abstract
AdaBoost.M1 has been successfully applied to improve the accuracy of a learning algorithm for multi-class classification problems. However, it assumes that the performance of each base classifier must be better than 1/2, and this may be hard to achieve in practice for a multi-class problem. A new algorithm called AdaBoost.MK only requiring base classifiers better than a random guessing (1/k) is thus designed.
Early SVM-based multi-class classification algorithms work by splitting the original problem into a set of two-class sub-problems. The time and space required by these algorithms are very demanding. In order to have low time and space complexities, we develop a base classifier that integrates one-class SVMs with discriminant functions.
In this study, a hybrid method that integrates AdaBoost.MK and one-class SVMs with improved discriminant functions as the base classifiers is proposed to solve a multi-class classification problem. Experimental results on data sets from UCI and Statlog show that the proposed approach outperforms many popular multi-class algorithms including support vector clustering and AdaBoost.M1 with one-class SVMs as the base classifiers.
目次 Table of Contents
摘要 i
Abstract ii
第一章 簡介 1
第二章 文獻探討 6
2.1 核心函式 6
2.2單類別支援向量機 10
2.3支援向量群聚 15
2.4支援向量機應用於多類別分類 18
第三章 研究方法 - KOCSVM & Adaboost.MK 22
3.1研究動機 22
3.2研究方法 - KOCSVM 27
3.2.1方法流程 27
3.2.2判別函式 31
3.3 Boosting演算法 34
3.3.1 Adaboost.M1演算法 35
3.3.2 Adaboost.MK演算法 38
第四章 實驗結果與分析 43
4.1資料集介紹 43
4.2實驗一 KOCSVM與KSVDD的比較 44
4.3實驗二 加入Adaboost.M1演算法前後之比較 48
4.4實驗三 Adaboost.MK與Adaboost.M1的比較 51
第五章 結論與未來研究方向 54
5.1 結論 54
5.2 未來研究方向 55
參考文獻 56
附錄一 支援向量機簡介 59
附錄二 序列最小優化簡介 61
附錄三 兩類別的Adaboost演算法 63
參考文獻 References
[1] K. Goh, E.Y. Chang, and B. Li, “Using One-class and Two-Class
Svms for Multiclass Image Annotation”, IEEE Transaction on
Knowledge and. Data Engineering, Vol. 17, No. 10, pp. 1333-1346,
2005.
[2] C.W. Hsu and C.J. Lin, “A comparison of methods for multi-class
support vector machines”, IEEE Transactions on Neural Networks,
Vol. 13, pp. 415-425, 2002.
[3] R. Rifin and A. Klautau, “In Defense of One-Vs-All Classification”,
Journal of Machine Learning Research, Vol. 5, pp. 101-141, 2004.
[4] Y. Freund and R. Schapire, “A decision-theoretic generalization of
on-line learning and an application to boosting”, Journal of
Computer and System Sciences, Vol. 55, No. 1, pp. 119-139, 1997.
[5] A. Ben-Hur, D. Horn, H.T. Siegelmann, and V. Vapnik. “Support
Vector Clustering”, Journal of Machine Learning Research, Vol. 2,
pp. 125-137, 2001.
[6] B.Y. Sun and D.S Huang, “Support vector clustering for multi-class
classification problems”, The Congress on Evolutionary
Computation, Canberra, Australia, pp.1480-1485, 2003.
[7] L. Manevitz and M. Yousef. “One-Class Svms for Document
Classification”, Journal of Machine Learning Research, Vol. 2, pp.
139-154, 2002.
[8] B. Schölkopf, J. Platt, J. Shawe-Taylor, A.J. Smola, and R.C.
Williamson, “Estimating the Support of a High-Dimensional
Distribution”, Neural Computation, Vol. 13, No. 7, pp. 1443-1472,
2001.
[9] Y. Freund and R. Schapire, “A decision-theoretic generalization of
on-line learning and an application to boosting”, Journal of
Computer and System Sciences, Vol. 55, No. 1, pp. 119-139, 1997.
[10] J. Shawe-Taylor and N. Cristianini, “Kernel Methods for Pattern
Analysis”, Cambridge,UK:Cambridge University Press,2004.
[11] K.-R. Muller, S. Mika, G. Ratsch, K. Tsuda, B. Scholkopf, “An
introduction to kernel-based learning algorithms”, IEEE
Transactions on Neural networks, Vol. 12, No. 2, pp. 181-201,
March 2001.
[12] L. Bottou, C. Cortes, J. Denker, H. Drucker, I. Guyon, L. D. Jackel,
Y. LeCun, U. Muller, E. Sackinger, P. Simard, and V. Vapnik,
“Comparison of classifier methods: a case study in handwritten digit
recognition”, In International Conference on Pattern Recognition,
pp. 77-87, 1994.
[13] J. Friedman, “Another Approach to Polychotomous Classification”,
Technical Report, Department of Statistics, Stanford University,
Stanford, CA. 1996.
[14] U. Kreßel, “Pairwise classification and support vector machines, In
B. Scholkopf, B.J. Burges, and A.J. Smola, editors, Advances in
Kernel Methods - Support Vector Learning”, pp. 255-268,
Cambridge, MA, MIT Press, 1999.
[15] J.C. Platt, N. Cristianini, and J. Shawe-Taylor, “Large margin DAGs
for multiclass classification”, In Advances in Neural Information
Processing System, Vol. 12, pp. 547-553, MIT Press, 2000.
[16] A. Sachs, C. Thiel and F. Schwenker, “One-Class Support-Vector
Machines for the Classification of Bioacoustic Time Series”,
International Journal on Artificial Intelligence and Machine
Learning, Vol. 6, No. 4, pp. 29-34, 2006.
[17] R. Schapire, “Using output codes to boost multiclass learning
problems”, In Proc. of the 14th International Conference on
Machine Learning, 1997.
[18] R. Schapire and Y. Singer, “Improved boosting algorithms using
confidence-rated prediction”, Machine Learning, Vol. 37, No. 1, pp.
297-336, 1999.
[19] L. Li, “Multiclass boosting with repartitioning”, In Proc. 23rd
International Conference on Machine Learning, 2006.
[20] G. Eibl, K.-P. Pfeiffer, “Multiclass Boosting for Weak Classifiers”,
Journal of Machine Learning Research, Vol. 6, No. Feb, pp.
189-210, 2005.
[21] J. Zhu, S. Rosset, H. Zhou, and T. Hastie, “Multiclass adaboost”,
Technical Report #430, Department of Statistics, University of
Michigan, 2005.
[22] C.C.Chang and C.J.Lin. LIBSVM: A Library for Support Vector
Machines. April 1, 2007.
[23] J.C.Platt, “Fast Training of Support Vector Machines using
Sequential Minimal Optimization”, Advances in Kernel Methods
-Support Vector Learning, pp. 185-208 , 1999.
[24] J.C.Platt, “Sequential Minimal Optimization: A Fast Algorithm for
Training Support Vector Machines”, Technical Report
MSR-TR-98-14, Microsoft Research, 1998.
電子全文 Fulltext
本電子全文僅授權使用者為學術研究之目的,進行個人非營利性質之檢索、閱讀、列印。請遵守中華民國著作權法之相關規定,切勿任意重製、散佈、改作、轉貼、播送,以免觸法。
論文使用權限 Thesis access permission:校內校外均不公開 not available
開放時間 Available:
校內 Campus:永不公開 not available
校外 Off-campus:永不公開 not available

您的 IP(校外) 位址是 3.138.141.202
論文開放下載的時間是 校外不公開

Your IP address is 3.138.141.202
This thesis will be available to you on Indicate off-campus access is not available.

紙本論文 Printed copies
紙本論文的公開資訊在102學年度以後相對較為完整。如果需要查詢101學年度以前的紙本論文公開資訊,請聯繫圖資處紙本論文服務櫃台。如有不便之處敬請見諒。
開放時間 available 已公開 available

QR Code