Boosting : foundations and algorithms / Robert E. Schapire and Yoav Freund.
نوع المادة : نصالسلاسل:Adaptive computation and machine learningالناشر:Cambridge (Mass.) : MIT press, [cop. 2012.]وصف:xv, 526 pages : illustrations ; 24 cmنوع المحتوى:- text
- unmediated
- volume
- 9780262017183
- 0262017180 (rel)
- Q325.75 .S33 2012
نوع المادة | المكتبة الحالية | رقم الطلب | رقم النسخة | حالة | تاريخ الإستحقاق | الباركود | |
---|---|---|---|---|---|---|---|
كتاب | UAE Federation Library | مكتبة اتحاد الإمارات General Collection | المجموعات العامة | Q325.75 .S33 2012 (إستعراض الرف(يفتح أدناه)) | C.1 | Library Use Only | داخل المكتبة فقط | 30010011136770 | ||
كتاب | UAE Federation Library | مكتبة اتحاد الإمارات General Collection | المجموعات العامة | Q325.75 .S33 2012 (إستعراض الرف(يفتح أدناه)) | C.2 | المتاح | 30010011136769 |
Browsing UAE Federation Library | مكتبة اتحاد الإمارات shelves, Shelving location: General Collection | المجموعات العامة إغلاق مستعرض الرف(يخفي مستعرض الرف)
Bibliogr. pages 501-510.
1. Introduction and Overview ; 1.1. Classification Problems and Machine Learning ; 1.2. Boosting ; 1.3. Resistance to Overfitting and the Margins Theory ; 1.4. Foundations and Algorithms ; Summary ; Bibliographic Notes ; Exercises ; I. Core Analysis ; 2. Foundations of Machine Learning ; 2.1. Direct Approach to Machine Learning ; 2.2. General Methods of Analysis ; 2.3. Foundation for the Study of Boosting Algorithms ; Summary ; Bibliographic Notes ; Exercises ; 3. Using AdaBoost to Minimize Training Error ; 3.1. Bound on AdaBoost's Training Error ; 3.2. Sufficient Condition for Weak Learnability ; 3.3. Relation to Chernoff Bounds ; 3.4. Using and Designing Base Learning Algorithms ; Summary ; Bibliographic Notes ; Exercises ; 4. Direct Bounds on the Generalization Error ; 4.1. Using VC Theory to Bound the Generalization Error ; 4.2. Compression-Based Bounds ; 4.3. Equivalence of Strong and Weak Learnability ; Summary ; Bibliographic Notes ; Exercises ; 5. Margins Explanation for Boosting's Effectiveness ; 5.1. Margin as a Measure of Confidence ; 5.2. Margins-Based Analysis of the Generalization Error ; 5.3. Analysis Based on Rademacher Complexity ; 5.4. Effect of Boosting on Margin Distributions ; 5.5. Bias, Variance, and Stability ; 5.6. Relation to Support-Vector Machines ; 5.7. Practical Applications of Margins ; Summary ; Bibliographic Notes ; Exercises ; II. Fundamental Perspectives ; 6. Game Theory, Online Learning, and Boosting ; 6.1. Game Theory ; 6.2. Learning in Repeated Game Playing ; 6.3. Online Prediction ; 6.4. Boosting ; 6.5. Application to a "Mind-Reading" Game ; Summary ; Bibliographic Notes ; Exercises ; 7. Loss Minimization and Generalizations of Boosting ; 7.1. AdaBoost's Loss Function ; 7.2. Coordinate Descent ; 7.3. Loss Minimization Cannot Explain Generalization ; 7.4. Functional Gradient Descent ; 7.5. Logistic Regression and Conditional Probabilities ; 7.6. Regularization ; 7.7. Applications to Data-Limited Learning ; Summary ; Bibliographic Notes ; Exercises ; 8. Boosting, Convex Optimization, and Information Geometry ; 8.1. Iterative Projection Algorithms ; 8.2. Proving the Convergence of AdaBoost ; 8.3. Unification with Logistic Regression ; 8.4. Application to Species Distribution Modeling ; Summary ; Bibliographic Notes ; Exercises ; III. Algorithmic Extensions ; 9. Using Confidence-Rated Weak Predictions ; 9.1. Framework ; 9.2. General Methods for Algorithm Design ; 9.3. Learning Rule-Sets ; 9.4. Alternating Decision Trees ; Summary ; Bibliographic Notes ; Exercises ; 10. Multiclass Classification Problems ; 10.1. Direct Extension to the Multiclass Case ; 10.2. One-against-All Reduction and Multi-label Classification ; 10.3. Application to Semantic Classification ; 10.4. General Reductions Using Output Codes ; Summary ; Bibliographic Notes ; Exercises ; 11. Learning to Rank ; 11.1. Formal Framework for Ranking Problems ; 11.2. Boosting Algorithm for the Ranking Task ; 11.3. Methods for Improving Efficiency ; 11.4. Multiclass, Multi-label Classification ; 11.5. Applications ; Summary ; Bibliographic Notes ; Exercises ; IV. Advanced Theory ; 12. Attaining the Best Possible Accuracy ; 12.1. Optimality in Classification and Risk Minimization ; 12.2. Approaching the Optimal Risk ; 12.3. How Minimizing Risk Can Lead to Poor Accuracy ; Summary ; Bibliographic Notes ; Exercises ; 13. Optimally Efficient Boosting ; 13.1. Boost-by-Majority Algorithm ; 13.2. Optimal Generalization Error ; 13.3. Relation to AdaBoost ; Summary ; Bibliographic Notes ; Exercises ; 14. Boosting in Continuous Time ; 14.1. Adaptiveness in the Limit of Continuous Time ; 14.2. BrownBoost ; 14.3. AdaBoost as a Special Case of BrownBoost ; 14.4. Experiments with Noisy Data ; Summary ; Bibliographic Notes ; Exercises ; Appendix: Some Notation, Definitions, and Mathematical Background ; A.1. General Notation ; A.2. Norms ; A.3. Maxima, Minima, Suprema, and Infima ; A.4. Limits ; A.5. Continuity, Closed Sets, and Compactness ; A.6. Derivatives, Gradients, and Taylor's Theorem ; A.7. Convexity ; A.8. Method of Lagrange Multipliers ; A.9. Some Distributions and the Central Limit Theorem.