帳號:guest(216.73.216.146)          離開系統
字體大小: 字級放大   字級縮小   預設字形  

詳目顯示

以作者查詢圖書館館藏以作者查詢臺灣博碩士論文系統以作者查詢全國書目
作者(中文):池伯遠
作者(外文):Chi, Bo-Yuan
論文名稱(中文):基於壓縮感知的縮減支持向量機研究
論文名稱(外文):A Study of Reduced Support Vector Machines based on Compressed Sensing
指導教授(中文):李育杰
張介玉
指導教授(外文):Lee, Yuh-Jye
Chang, Chieh-Yu
口試委員(中文):朱家杰
銀慶剛
陳素雲
口試委員(外文):Chu, Chia-Chieh
Ing, Ching-Kang
Huang, Su-Yuan
學位類別:碩士
校院名稱:國立清華大學
系所名稱:數學系
學號:104021614
出版年(民國):106
畢業學年度:106
語文別:英文
論文頁數:43
中文關鍵詞:支持向量機縮減支持向量機
外文關鍵詞:SVMRSVM
相關次數:
  • 推薦推薦:0
  • 點閱點閱:981
  • 評分評分:*****
  • 下載下載:0
  • 收藏收藏:0
為了克服大規模分類問題中生成非線性支持向量機的計算負擔,提出了縮減支持向量機。縮減內核技巧已經成功的被應用在許多基於核方法的機器學習演算法。在本論文中,我們嘗試將隨機選擇小的縮減集與隨機投影聯繫起來。這對縮減支持向量機提供了基於壓縮學習的新解釋。我們利用稀疏編碼在高維特徵空間中表示核向量,並將這些稀疏向量隨機投影到較低維度的壓縮域中。我們的實驗結果表明,在高維空間中的全核非線性支持向量機、縮減支持向量機、在高維特徵空間中的線性支持向量機以及在壓縮空間中的線性支持向量機,都具有相似的分類表現。我們同時發現,解決壓縮空間中的機器學習問題比恢復稀疏信號要容易的多。這意味著你可以更積極的使用更小的縮減集來生成分類器。然而,如果要從壓縮信號來重建原始的稀疏向量,壓縮空間的維度必須要大於縮減集的大小。
The reduced support vector machine was proposed for the practical objective to overcome the computational burden in generating a nonlinear SVM for the large-scale classification problems. The reduced (rectangle) kernel trick has been applied to many machine learning algorithms in cooperating with kernel functions. In this thesis, we try to link the random selection small reduced set to random projection. It provides a new interpretation of RSVM via a compressed learning viewpoint. We utilized sparse coding to represent the kernel vectors in a high dimensional feature space and random project these sparse coded vectors into a lower dimensional compressed domain. Our empirical results show that the nonlinear SVM with full kernel, RSVM, linear SVM in the high dimensional feature space and linear SVM in the compressed space all of them have a very similar performance. We also find that solving a machine learning in compressed space is much easier than recovering a sparse signal. That means you can be more aggressively using a small reduced set in generating a classifier. However, if you want to reconstruct the original sparse vector from the compressed signal than the dimension of compressed space has to bigger than the size of reduced set.
1 Introduction 1
2 Related works 4
2.1 SVM, SSVM, and RSVM 4
2.2 Compressed Sensing and Learning 10
2.3 Sparse Coding and Dictionary Learning 17
3 Our Method 22
3.1 Four Layers Models 23
3.2 Connect to RSVM 28
4 Numerical results 32
5 Conclusion 38
Reference 40
[1] Michal Aharon, Michael Elad, and Alfred Bruckstein. k-svd: An algorithm for designing overcomplete dictionaries for sparse representation. IEEE Transactions on Signal Processing, 54(11):4311–4322, 2006.
[2] Chenglong Bao, Jian-Feng Cai, and Hui Ji. Fast sparsity-based orthogonal dictionary learning for image restoration. In Proceedings of the IEEE International Conference on Computer Vision, pages 3384–3391, 2013.
[3] Richard Baraniuk, Mark Davenport, Ronald DeVore, and Michael Wakin. A simple proof of the restricted isometry property for random matrices. Constructive Approximation, 28(3):253–263, 2008.
[4] Alfred M Bruckstein, David L Donoho, and Michael Elad. From sparse solutions of systems of equations to sparse modeling of signals and images. SIAM Review, 51(1):34–81, 2009.
[5] Kurt Bryan and Tanya Leise. Making do with less: an introduction to compressed sensing. SIAM Review, 55(3):547–566, 2013.
[6] Christopher JC Burges. A tutorial on support vector machines for pattern recognition. Data Mining and Knowledge Discovery, 2(2):121–167, 1998.
[7] Robert Calderbank and Sina Jafarpour. Finding needles in compressed haystacks. In Acoustics, Speech and Signal Processing (ICASSP), 2012 IEEE International Conference on, pages 3441–3444. IEEE, 2012.
[8] Robert Calderbank, Sina Jafarpour, and Robert Schapire. Compressed learning: Universal sparse dimensionality reduction and learning in the measurement domain. preprint, 2009.
[9] Emmanuel J Candes and Terence Tao. Decoding by linear programming. IEEE Transactions on Information Theory, 51(12):4203–4215, 2005.
[10] Chien-Chung Chang and Yuh-Jye Lee. Generating the reduced set by systematic sampling. In IDEAL, pages 720–725. Springer, 2004.
[11] Li-Jen Chien, Chien-Chung Chang, Yuh-Jye Lee, et al. Variant methods of reduced set selection for reduced support vector machines. J. Inf. Sci. Eng., 26(1):183–196, 2010.
[12] Adam Coates and Andrew Y Ng. Learning feature representations with k-means. In Neural Networks: Tricks of the Trade, pages 561–580. Springer, 2012.
[13] Sanjoy Dasgupta and Anupam Gupta. An elementary proof of a theorem of johnson and lindenstrauss. Random Structures Algorithms, 22(1):60–65,
2003.
[14] David L Donoho. Compressed sensing. IEEE Transactions on Information Theory, 52(4):1289–1306, 2006.
[15] David L Donoho and Michael Elad. Optimally sparse representation in general (nonorthogonal) dictionaries via l1 minimization. Proceedings of the National Academy of Sciences, 100(5):2197–2202, 2003.
[16] Bradley Efron, Trevor Hastie, Iain Johnstone, Robert Tibshirani, et al. Least angle regression. The Annals of statistics, 32(2):407–499, 2004.
[17] Kjersti Engan, Sven Ole Aase, and J Hakon Husoy. Method of optimal directions for frame design. In Acoustics, Speech, and Signal Processing, 1999. Proceedings., 1999 IEEE International Conference on, volume 5, pages 2443–2446. IEEE, 1999.
[18] Elaine T Hale, Wotao Yin, and Yin Zhang. A fixed-point continuation method for l1-regularized minimization with applications to compressed sensing. CAAM TR07-07, Rice University, 43:44, 2007.
[19] Lih-Ren Jen and Yuh-Jye Lee. Clustering model selection for reduced support vector machines. In IDEAL, pages 714–719. Springer, 2004.
[20] Honglak Lee, Alexis Battle, Rajat Raina, and Andrew Y Ng. Efficient sparse coding algorithms. In Advances in Neural Information Processing Systems, pages 801–808, 2007.
[21] Yuh-Jye Lee and Su-Yun Huang. Reduced support vector machines: A statistical theory. IEEE Transactions on Neural Networks, 18(1):1–13, 2007.
[22] Yuh-Jye Lee, Hung-Yi Lo, and Su-Yun Huang. Incremental reduced support vector machines. In International Conference on Informatics Cybernetics and System (ICICS 2003), 2003.
[23] Yuh-Jye Lee and Olvi L Mangasarian. Rsvm: Reduced support vector machines. In Proceedings of the 2001 SIAM International Conference on Data Mining, pages 1–17. SIAM, 2001.
[24] Yuh-Jye Lee and Olvi L Mangasarian. Ssvm: A smooth support vector machine for classification. Computational optimization and Applications, 20(1):5–22, 2001.
[25] Yuh-Jye Lee, Yi-Ren Yeh, and Hsing-Kuo Pao. Introduction to support vector machines and their applications in bankruptcy prognosis. In Handbook of Computational Finance, pages 731–761. Springer, 2012.
[26] Kuan-Ming Lin and Chih-Jen Lin. A study on reduced support vector machines. IEEE Transactions on Neural Networks, 14(6):1449–1459, 2003.
[27] Julien Mairal, Francis Bach, and Jean Ponce. Task-driven dictionary learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 34(4):791–804, 2012.
[28] Julien Mairal, Francis Bach, Jean Ponce, and Guillermo Sapiro. Online dictionary learning for sparse coding. In Proceedings of the 26th Annual International Conference on Machine Learning, pages 689–696. ACM, 2009.
[29] Julien Mairal, Francis Bach, Jean Ponce, and Guillermo Sapiro. Online learning for matrix factorization and sparse coding. Journal of Machine Learning Research, 11(Jan):19–60, 2010.
[30] Julien Mairal, Francis Bach, Jean Ponce, Guillermo Sapiro, and Andrew Zisserman. Discriminative learned dictionaries for local image analysis. In Computer Vision and Pattern Recognition, 2008. CVPR 2008. IEEE Conference on, pages 1–8. IEEE, 2008.
[31] Stéphane Mallat. A wavelet tour of signal processing. Academic press, 1999.
[32] Olvi L Mangasarian. Generalized support vector machines. Advances in Neural Information Processing Systems, pages 135–146, 1999.
[33] Bruno A Olshausen and David J Field. Sparse coding with an overcomplete basis set: A strategy employed by v1. Vision research, 37(23):3311–3325, 1997.
[34] Rajat Raina, Alexis Battle, Honglak Lee, Benjamin Packer, and Andrew Y Ng. Self-taught learning: transfer learning from unlabeled data. In Proceedings of the 24th Annual International Conference on Machine Learning, pages 759–766. ACM, 2007.
[35] Alex J Smola and Bernhard Schölkopf. Learning with kernels. GMDForschungszentrum Informationstechnik, 1998.
[36] Robert Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society. Series B (Methodological), pages 267–288, 1996.
[37] Joel A Tropp and Anna C Gilbert. Signal recovery from random measurements via orthogonal matching pursuit. IEEE Transactions on Information Theory, 53(12):4655–4666, 2007.
[38] Joel A Tropp and Stephen J Wright. Computational methods for sparse solution of linear inverse problems. Proceedings of the IEEE, 98(6):948–958, 2010.
[39] Vladimir Vapnik. The nature of statistical learning theory. Springer science business media, 2013.
[40] Christopher KI Williams and Matthias Seeger. Using the nyström method to speed up kernel machines. In Advances in Neural information Processing Systems, pages 682–688, 2001.
[41] John Wright, Allen Y Yang, Arvind Ganesh, S Shankar Sastry, and Yi Ma. Robust face recognition via sparse representation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 31(2):210–227, 2009. 43
(此全文未開放授權)
電子全文
中英文摘要
 
 
 
 
第一頁 上一頁 下一頁 最後一頁 top
* *