Does anyone know whether one could use Adaboost with random forest base classifiers? I searched the web to learn more about this, but most webpages provided comparisons of the two as ensemble learning methods, and I didn't find anything about using them together.
(Basically somebody asked it here, but the only answer so far contradicts with my observations, which I'm sharing below)
Notwithstanding, nobody explicitly said there was anything wrong with it, so I tried it on a typical dataset with n rows of p real-valued features, and a label list of length n. In case it matters, they are embeddings of nodes in a graph obtained by the DeepWalk algorithm, and the nodes are categorized into two classes. I trained a few classification models on this data using 5-fold cross validation, and measured common evaluation metrics for them (precision, recall, AUC etc.). The models I have used are SVM, logistic regression, random Forest, 2-layer perceptron and Adaboost with random forest classifiers. The last model, Adaboost with random forest classifiers, yielded the best results. Sure, now the runtime has increased by a factor of, let's say, 100, but it's still about 20 mins, so it's not a constraint to me. Now I wonder if I should be suspicious about its good accuracy (95% AUC compared to multilayer perceptron's 89% and random forest's 88%)
Here's what I thought: Firstly, I'm using cross validation, so there's probably no overfitting flying under the radar. Secondly, both are ensemble learning methods, but random forest is a bagging method, wheras Adaboost is a boosting technique. So they're still different enough for their combination to make sense.
Am I wrong?