[R] randomForest question

From: <Arne.Muller_at_sanofi-aventis.com>
Date: Wed 26 Jul 2006 - 21:32:52 EST


I've a question regarding randomForest (from the package with same name). I've 16 featurs (nominative), 159 positive and 318 negative cases that I'd like to classify (binary classification).

Using the tuning from the e1071 package it turns out that the best performance if reached when using all 16 features per tree (mtry=16). However, the documentation of randomForest suggests to take the sqrt(#features), i.e. 4. How can I explain this difference? When using all features this is the same as a classical decision tree, with the difference that the tree is built and tested with different data sets, right?

example (I've tried different configurations, incl. changing ntree):
> param <- try(tune(randomForest, class ~ ., data=d.all318, range=list(mtry=c(4, 8, 16), ntree=c(1000))));
> summary(param)

Parameter tuning of `randomForest':

        thanks a lot for your help,

        kind regards,

R-help@stat.math.ethz.ch mailing list
https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code. Received on Wed Jul 26 21:38:41 2006

Archive maintained by Robert King, hosted by the discipline of statistics at the University of Newcastle, Australia.
Archive generated by hypermail 2.1.8, at Wed 26 Jul 2006 - 23:03:45 EST.

Mailing list information is available at https://stat.ethz.ch/mailman/listinfo/r-help. Please read the posting guide before posting to the list.