Remember Me
Or use your Academic/Social account:


Or use your Academic/Social account:


You have just completed your registration at OpenAire.

Before you can login to the site, you will need to activate your account. An e-mail will be sent to you with the proper instructions.


Please note that this site is currently undergoing Beta testing.
Any new content you create is not guaranteed to be present to the final version of the site upon release.

Thank you for your patience,
OpenAire Dev Team.

Close This Message


Verify Password:
Verify E-mail:
*All Fields Are Required.
Please Verify You Are Human:
fbtwitterlinkedinvimeoflicker grey 14rssslideshare1
Palczewska, A; Palczewski, J; Robinson, RM; Neagu, D (2013)
Publisher: IEEE
Languages: English
Types: Other
Model interpretation is one of the key aspects of the model evaluation process. The explanation of the relationship between model variables and outputs is easy for statistical models, such as linear regressions, thanks to the availability of model parameters and their statistical significance. For “black box” models, such as random forest, this information is hidden inside the model structure. This work presents an approach for computing feature contributions for random forest classification models. It allows for the determination of the influence of each variable on the model prediction for an individual instance. Interpretation of feature contributions for two UCI benchmark datasets shows the potential of the proposed methodology. The robustness of results is demonstrated through an extensive analysis of feature contributions calculated for a large number of generated random forest models.
  • The results below are discovered through our pilot algorithms. Let us know how we are doing!

    • [1] Breast Cancer Wisconsin Diagnostic dataset. http:// archive.ics.uci.edu/ml/datasets/Breast+ Cancer+Wisconsin+\%28Diagnostic\%29.
    • [2] CRAN - The Comprehensive R Archive Network. http: //cran.r-project.org/.
    • [3] Iris dataset. http://archive.ics.uci.edu/ml/ datasets/Iris.
    • [4] D. Baehrens, T. Schroeter, S. Harmeling, M. Kawanabe, K. Hansen, and K.-R. Muller. How to explain individual classification decisions. Journal of Machine Learning Research, 11:1803-1831, 2010.
    • [5] L. Breiman. Random forests. Machine Learning, 45(1):5- 32, 2001.
    • [6] L. Breiman and A. Cutler. Random forests. http://www.stat.berkeley.edu/˜breiman/ RandomForests/, 2008.
    • [7] L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stone. Classification and regression trees. Monterey, CA: Wadsworth & Brooks/Cole Advanced Books & Software, 1984.
    • [8] K. Hansen, D. Baehrens, T. Schroeter, M. Rupp, and K.- R. Mller. Visual interpretation of kernel-based prediction models. Molecular Informatics, 30(9):817-826, 2011.
    • [9] V. E. Kuz'min, P. G. Polishchuk, A. G. Artemenko, and S. A. Andronati. Interpretation of qsar models based on random forest methods. Molecular Informatics, 30(6-7):593-603, 2011.
    • [10] A. Liaw and M. Wiener. Classification and regression by randomforest. R News, 2(3):18-22, 2002.
    • [11] L. Rosenbaum, G. Hinselmann, A. Jahn, and A. Zell. Interpreting linear support vector machine models with heat map molecule coloring. Journal of Cheminformatics, 3(1):11, 2011.
  • No related research data.
  • No similar publications.

Share - Bookmark

Cite this article