Book contents
- Frontmatter
- Contents
- Preface
- Acronyms
- 1 Introduction
- 2 Machine Learning and Statistics Overview
- 3 Performance Measures I
- 4 Performance Measures II
- 5 Error Estimation
- 6 Statistical Significance Testing
- 7 Datasets and Experimental Framework
- 8 Recent Developments
- 9 Conclusion
- Appendix A Statistical Tables
- Appendix B Additional Information on the Data
- Appendix C Two Case Studies
- Bibliography
- Index
8 - Recent Developments
Published online by Cambridge University Press: 05 August 2011
- Frontmatter
- Contents
- Preface
- Acronyms
- 1 Introduction
- 2 Machine Learning and Statistics Overview
- 3 Performance Measures I
- 4 Performance Measures II
- 5 Error Estimation
- 6 Statistical Significance Testing
- 7 Datasets and Experimental Framework
- 8 Recent Developments
- 9 Conclusion
- Appendix A Statistical Tables
- Appendix B Additional Information on the Data
- Appendix C Two Case Studies
- Bibliography
- Index
Summary
We reviewed the major components of the evaluation framework in the last chapters and described in detail the various techniques pertaining to each, together with their applications. The focus of this chapter is to complement this review by outlining various advancements that have been made relatively recently, but that have not yet become mainstream. We also look into some approaches aimed at addressing problems arising from the developments on the machine learning front in various application settings, such as ensemble classifiers.
Just as with the traditional developments in performance evaluation, recent attempts at improving as well as designing new performance metrics have led the way. These have resulted in both improvements to existing performance metrics, thereby claiming to ameliorate the issues with the current versions, and proposals for novel metrics aimed at addressing the areas of algorithm evaluation not satisfactorily addressed by current metrics. We discuss in brief some of these advancements in Section 8.1. In Section 8.2, we focus on the attempts at unifying these performance metrics as well as studying their interrelation in the form of both theoretical and experimental frameworks. A natural extension to such studies is the design of more general or broader measures of performance, either as a result of insights obtained from the theoretical framework or by combining existing metrics based on observations from the experimental frameworks. Such metric combinations for obtaining general measures are the focus of Section 8.3.
- Type
- Chapter
- Information
- Evaluating Learning AlgorithmsA Classification Perspective, pp. 308 - 334Publisher: Cambridge University PressPrint publication year: 2011