好文档就是一把金锄头!
欢迎来到金锄头文库![会员中心]
电子文档交易市场
安卓APP | ios版本
电子文档交易市场
安卓APP | ios版本

cop5992–dataminingtermprojectrandom…:cop5992–数据挖掘项目随机….ppt

13页
  • 卖家[上传人]:tian****1990
  • 文档编号:81414982
  • 上传时间:2019-02-21
  • 文档格式:PPT
  • 文档大小:312.31KB
  • / 13 举报 版权申诉 马上下载
  • 文本预览
  • 下载提示
  • 常见问题
    • COP5992 – DATA MINING TERM PROJECT RANDOM SUBSPACE METHOD + CO-TRAINING by SELIM KALAYCI,RANDOM SUBSPACE METHOD (RSM),Proposed by Ho “The Random Subspace for Constructing Decision Forests”, 1998 Another combining technique for weak classifiers like Bagging, Boosting.,RSM ALGORITHM,1. Repeat for b = 1, 2, . . ., B: (a) Select an r-dimensional random subspace X from the original p-dimensional feature space X. 2. Combine classifiers Cb(x), b = 1, 2, . . ., B, by simple majority voting to a final decision rule,MOTIVATION FOR RSM,Redundancy in Data Feature Space Completely redundant feature set Redundancy is spread over many features Weak classifiers that have critical training sample sizes,RSM PERFORMANCE ISSUES,RSM Performance depends on: Training sample size The choice of a base classifier The choice of combining rule (simple majority vs. weighted) The degree of redundancy of the dataset The number of features chosen,DECISION FORESTS (by Ho),A combination of trees instead of a single tree Assumption: Dataset has some redundant features Works efficiently with any decision tree algorithm and data splitting method Ideally, look for best individual trees with lowest tree similarity,UNLABELED DATA,Small number of labeled documents Large pool of unlabeled documents How to classify unlabeled documents accurately?,EXPECTATION-MAXIMIZATION (E-M),CO-TRAINING,Blum and Mitchel, “Combining Labeled and Unlabeled Data with Co-Training”, 1998. Requirements: Two sufficiently strong feature sets Conditionally independent,CO-TRAINING,APPLICATION OF CO-TRAINING TO A SINGLE FEATURE SET,Algorithm: Obtain a small set L of labeled examples Obtain a large set U of unlabeled examples Obtain two sets F1 and F2 of features that are sufficiently redundant While U is not empty do: Learn classifier C1 from L based on F1 Learn classifier C2 from L based on F2 For each classifier Ci do: Ci labels examples from U based on Fi Ci chooses the most confidently predicted examples E from U E is removed from U and added (with their given labels) to L End loop,THINGS TO DO,How can we measure redundancy and use it efficiently? Can we improve Co-training? How can we apply RSM efficiently to: Supervised learning Semi-supervised learning Unsupervised learning,QUESTIONS,????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????,。

      点击阅读更多内容
      关于金锄头网 - 版权申诉 - 免责声明 - 诚邀英才 - 联系我们
      手机版 | 川公网安备 51140202000112号 | 经营许可证(蜀ICP备13022795号)
      ©2008-2016 by Sichuan Goldhoe Inc. All Rights Reserved.