https://scholars.lib.ntu.edu.tw/handle/123456789/81925
DC 欄位 | 值 | 語言 |
---|---|---|
dc.contributor | 陳正剛 | en |
dc.contributor | 臺灣大學:工業工程學研究所 | zh_TW |
dc.contributor.author | 徐鍾睿 | zh |
dc.contributor.author | Hsu, Chung-Jui | en |
dc.creator | 徐鍾睿 | zh |
dc.creator | Hsu, Chung-Jui | en |
dc.date | 2006 | en |
dc.date.accessioned | 2007-11-26T01:07:56Z | - |
dc.date.accessioned | 2018-06-29T00:33:34Z | - |
dc.date.available | 2007-11-26T01:07:56Z | - |
dc.date.available | 2018-06-29T00:33:34Z | - |
dc.date.issued | 2006 | - |
dc.identifier | en-US | en |
dc.identifier.uri | http://ntur.lib.ntu.edu.tw//handle/246246/51215 | - |
dc.description.abstract | 屬性解釋(Attribute Interpretation)對於分類方法是非常重要的,但一般非線性的分類方法,都無法提供屬性解釋的功能。因此,這篇論文的目的,是希望能夠建立一個方法,讓核心費雪區別法(Kernel Fisher Discriminant)和核心最小平方誤差法(Kernel Minimum Squared Error)兩種非線性的核分類方法亦能提供屬性解釋的能力。這兩種方法都是把實例(instances)從原本的屬性空間(attribute space)投射到高維度的特徵空間(feature space)裡,並在特徵空間裡有效的對實例做樣式辨認(pattern recognition)或特徵提取(feature extraction),卻也因此失去了屬性解釋的功能。因此,我們可以利用核區別得點(kernel discriminant score)將原本的實例切割成幾個具有線性結構的小組,並在每個小組中使用線性的費雪區別法(Fisher Linear Discriminant),如此便可得到每個小組中的屬性解釋。我們所提出的方法稱為核分組費雪區別法(Kernel-Partition FLD)。此外,我們也將針對核分組來提供屬性解釋。我們將證明核分組的分類方法不但可以如同非線性的分類方法有較好的區分能力,另外也可以針對每個小組提供屬性解釋的功能,並且知道每一個組別中重要的屬性為何。我們也將核分組費雪區別法應用在文件分類的問題中。文件資料的高維度造成了運算的成本高昂,並且可能存在許多不必要的雜訊影響分類的結果。因此,我們也提出了一個減少維度的方法,並證明這個方法不但減輕了我們的運算成本並且將因減少維度而損失的資訊降到最低。對於核分組費雪區別方法,這些屬性解釋甚至可以成為核分類方法的分類依據,我們先在特徵空間裡利用適當的區別得點將實例分組,然後建構出區別函數使得預測的正確率可以達到最高。最後經由一些模擬和實際的文件分類之例子證實,核分組費雪區別法確實結合了一般線性和非線性分類方法的優點。 | zh_TW |
dc.description.abstract | Since attribute interpretation is important in classification but not provided by nonlinear classifiers, the objective of this research is to develop a methodology for nonlinear classification methods, namely KFD (Kernel Fisher Discriminant) and KMSE (Kernel Minimum Squared Error), to provide attribute interpretation. The proposed methodology is called Kernel-Partition FLD. KFD and KMSE are both kernel-based classifiers which transform the instances from the original attribute space to the feature space. The feature space is efficient for feature extraction and pattern recognition but losses the meanings of the original attributes. For attribute interpretation, we need to partition the instances with nonlinear structures into several groups where each group has its own linear structure. Then, we can apply FLD (Fisher Linear Discriminant) for each group to provide attribute interpretation. In addition, we also attempt to attribute interpretation for the Kernel-Partition in this study. We will then apply the methodology to document classification. Classification of a large number of documents with a great number of terms is a challenge for all learning algorithms and will be the focus of this research. A novel approach should be developed such that the text dataset can be better classified through nonlinear classification with knowledge on which terms (attributes) are more important for classification of certain types of documents. Moreover, the high computation cost and the sparsity problem of document vectors in document classification is also an issue to be addressed in this research. Thus, a dimension-reduction methodology is developed to effectively diminish the computation requirement and reduce the dimensions without loss much information. With Kernel-Partition FLD, the attribute interpretation can be further developed to become classification rules for kernel-based classification approaches. In this research, the proposed methodologies will be shown to successfully combine the advantages of both linear and nonlinear classifiers through simulated cases and the real-world cases of the text dataset. | en |
dc.description.tableofcontents | Abstract i 論文摘要 ii Contents iii Contents of Figures v Contents of Tables vi Chapter 1: Introduction 1 1.1 Background 1 1.2 Document Classification 2 1.3 Current Classification Approaches and Problems of Current Nonlinear Classification Approaches 3 1.3.1 Principal Component Analysis (PCA) 3 1.3.2 Fisher Linear Discriminant (FLD) 5 1.4 Problem Definition and Research Objective 9 1.5 Thesis Organization 10 Chapter 2: Kernel-Partition for Kernel-Based Classification 11 2.1 Kernel-Based Classification 11 2.1.1 Kernel Fisher Discriminant (KFD) 11 2.1.2 Kernel Minimum Squared Error (KMSE) 15 2.2 Attribute Interpretation and Classification Rules for Kernel-Based Classification 21 2.3 Illustration with simulated sample 30 2.3.1 Simulated Case 1: Double Ring 30 2.3.2 Simulated Case 2: Five Clusters 36 2.4 Dimensionality Reduction 41 Chapter 3: Case Study 48 3.1 Document Preprocessing and Representation 48 3.2 Medline Text Dataset 51 3.3 Med Text Dataset 65 Chapter 4: Conclusions and Future Research 70 References 72 Appendix A: 30 Classes of Med Text Dataset 76 Appendix B: User Interface 78 | zh_TW |
dc.format.extent | 961633 bytes | - |
dc.format.mimetype | application/pdf | - |
dc.language | en-US | en |
dc.language.iso | en_US | - |
dc.subject | 核費雪區別法 | en |
dc.subject | 核最小平方誤差法 | en |
dc.subject | 核分組費雪區別法 | en |
dc.subject | 屬性解釋 | en |
dc.subject | Document Classification | en |
dc.subject | Attribute Interpretation | en |
dc.subject | Kernel-Partition FLD | en |
dc.subject | Kernel Fisher discriminants | en |
dc.subject | Kernel Minimum Squared Error | en |
dc.title | 核分組費雪區別法於文件分類之應用 | zh_TW |
dc.title | Applications of Kernel-Partition FLD to Document Classification | en |
dc.type | thesis | en |
dc.identifier.uri.fulltext | http://ntur.lib.ntu.edu.tw/bitstream/246246/51215/1/ntu-95-R93546026-1.pdf | - |
dc.relation.reference | [1] Masand, B., G. Linoff, and D.Waltz, Classifying news stories using memory based reasoning, 15th Ann Int ACM SIGIR Conference on Research and Development in Information Retrieval , pp. 59–64, 1992. [2] Joachims, T., Text categorization with support vector machines: Learning withmany relevant features, European Conference on Machine Learning (ECML), 1998. [3] Wiener, E., J. O. Pedersen, and A. S. Weigend, “A neural network approach totopic spotting”, Proceedings of the Fourth Annual Symposium on Document Analysis and Information Retrieval , 1995. [4] Yang, Y. and X. Liu, A re-examination of text categorization methods, Proceedings of SIGIR-99, 22nd ACM International Conference on Research and Development in Information Retrieval , Berkeley, US, 1999. [5] McCallum, A. and K. Nigam, A comparison of event models for naive bayes text classification, AAAI-98 Workshop on Learning for Text Categorization, 1998. [6] I. Jollife. Principal Component Analysis. Spring-Verlag, New York,1986. [7] Fisher, R. A. (1936). The use of multiple measurements in taxonomic problems. Annals of Eugenics 7, 179-188. [8] Harry Hochstadt. Integral Equations. Wiley Classics Library. John Wiley & Sons Inc., New York, 1989. ISBN 0-471-50404-1. Reprint of the 1973 original, A Wiley-Interscience Publication. [9] B. Schölkopf, C. J. C. Burges, & A. J. Smola, editors. Advanced in Kernel Methods – Support Vector Machine. MIT Press, Cambridge, MA, 1999. [10] B. Schölkopf, A. Smola, & K-R Müller. Kernel principal component analysis. In B. Schölkopf, C. J. C. Burges, & A. J. Smola, editors, Advanced in Kernel Methods – Support Vector Machine. 327-352. MIT Press, Cambridge, MA, 1999. [11] Mika, S., Rätsch, G., Weston, J., Schölkopf, B., & Müller, K-R. (1999). Fisher Discriminant Analysis With Kernels. IEEE International Workshop on Neural Networks for Signal Processing, Vol. IX, Madison, USA, August 1999, 41-48. [12] Mercer kernels S. Saitoh, Theory of Reproducing Kernels and its Applications, Longman Scientific & Technical, Harlow, England, 1988. [13] B. Schölkopf, A. Smola, & K-R Müller. Nonlinear component analysis as kernel eigenvalue problem. Neural Computation, 10: 1299-1319, 1998. 2. [14] R.O.Duda and P.E. Hart. Pattern Classification and Scene Analysis. John Wiley & Sons, New York, 1973. [15] Z.Bian, X.Zhang, et al. Pattern Recognition (2nd ed., in Chinese). Tsinghua University Press, Beijing, 2000. [16] J.T.Tou and R.C. Gonzadez. Pattern Recognition Principles. Addison-Wesly, Reading, 1974. [17] S.Theodoridis and K. Koutroumbas. Pattern Recognition. Academic Press, San Diego, 1999. [18] S. Mika, G. Rätsch, and K.-R. Müller. A mathematical programming approach to the Kernel Fisher algorithm. In Advances in Neural Information Processing Systems 13, 2001. to appear. [19] Jianhua Xu, Xuegong Zhang, Yanda Li. Kernel MSE algorithm: a unified framework for KFD, LS-SVM and KRR. Proceedings of 2001 International Joint Conference on Neural Networks (IJCNN’01), 1486-1491, 2001. Washington DC, USA, July 15-19, 2001. (EI) [20] G. Strang. Linear Algebra and Its Applications. 1988. Thomson Learning, Inc., 3rd edition. [21] H. Kim and P. Howland. Dimension reduction in text classification using support vector machines, Journal of Machine Learning Research, 6, pp. 37-53, 2005 [22] Salton, G., C. Yang, and A. Wong, A vector-space model for automatic indexing,97 Communications of the ACM, Vol. 18, No. 11, pp. 613–620, 1975. [23] Ashish, Sen, and Muni Srivastava. Regression Analysis: theory, methods and applications. New York: Springer-Verlag, c1990. [24] M.H. DeGroot and M.J. Schervish, Probability and Statistics, third ed. Addison-Wesley, 2002. [25] Dumais, S., T. Platt, D. Heckermann, and M. Sahami, Inductive learning algorithms and representations for text categorization, Proceedings of the Seventh International Conference on Information and Knowledge Management, 1998. [26] Porter, M. F., An algorithm for suffix stripping, Program, Vol. 14, pp. 130–137, 1980. [27] ¨Ozg¨ur , Arzucan., Supervised and un supervised machine learning techniques for text document categorization, B.S. in Computer Engineering, Bo˘gazi¸ci University, 2002 [28] C. Elkan., Deriving TF-IDF as a Fisher kernel. To appear in Proceedings of the International Symposium on String Processing and Information Retrieval (SPIRE'05), Buenos Aires, Argentina, November 2005, pp. 296-301. [29] University of Minnesota. 2006. Computer Science & Engineering. 12 Jun. 2006 http://www-users.cs.umn.edu/~hpark/data.html. [30] M. W. Berry, S. T. Dumais, and G. W. O’Brien. Using linear algebra for intelligent information retrieval. SIAM Review, 37:573-595, 1995. [31] Latent Semantic Indexing Web Site. Common IR Test Collections. http://www.cs.utk.edu/~lsi/. | en |
item.openairecristype | http://purl.org/coar/resource_type/c_46ec | - |
item.openairetype | thesis | - |
item.languageiso639-1 | en_US | - |
item.grantfulltext | open | - |
item.cerifentitytype | Publications | - |
item.fulltext | with fulltext | - |
顯示於: | 工業工程學研究所 |
檔案 | 描述 | 大小 | 格式 | |
---|---|---|---|---|
ntu-95-R93546026-1.pdf | 23.53 kB | Adobe PDF | 檢視/開啟 |
在 IR 系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。