The Altair Community is migrating to a new platform to provide a better experience for you. In preparation for the migration, the Altair Community is on read-only mode from October 28 - November 6, 2024. Technical support via cases will continue to work as is. For any urgent requests from Students/Faculty members, please submit the form linked here
guidelines for base learner vs stack learner
Dear all, Im working on a classification problem with supervised learning, for which
I want to use the stacking operator to improve accuracy
1. while choosing stacking operator, im looking for Is there any guideline how to combine the operators.
Using Gradient boost alone gives accuracy 75.75%. I want to take it beyond 95% using stacking.
(ofcourse I'll look into the requirements of precision/recall).
I tried many stacking combinations, i couldnt make it beyond 73%.
2. Any theoretical/white paper reference / rapidminer resource/ book on how to choose the combination of base learners , stack learner for the maximum classification performance.
3. will it be useful to use non-neural net operator as base learner & neural net (deep learning ) as stacking learner?
i tried this combination, in my case, it reduces the performance to 55%.
thank you
thiru
I want to use the stacking operator to improve accuracy
1. while choosing stacking operator, im looking for Is there any guideline how to combine the operators.
Using Gradient boost alone gives accuracy 75.75%. I want to take it beyond 95% using stacking.
(ofcourse I'll look into the requirements of precision/recall).
I tried many stacking combinations, i couldnt make it beyond 73%.
2. Any theoretical/white paper reference / rapidminer resource/ book on how to choose the combination of base learners , stack learner for the maximum classification performance.
3. will it be useful to use non-neural net operator as base learner & neural net (deep learning ) as stacking learner?
i tried this combination, in my case, it reduces the performance to 55%.
thank you
thiru
0
Best Answer
-
Telcontar120 RapidMiner Certified Analyst, RapidMiner Certified Expert, Member Posts: 1,635 UnicornThere is no way to know in advance (unless you are dealing with an artificial dataset used specifically for instructional purposes) whether an accuracy of 95% is feasible on your data (without overfitting).
Stacking is a kind of ensemble modeling. So is Gradient Boosting. They are different approaches but Gradient Boosting "alone" is still an ensemble method. In theory you may be able to improve your results by putting GBT into your ensemble as the base learner and then using something else in combination with it to help with those cases where it it not doing a good job, but again, it might not significantly improve the results on your validation set since further improvements may not be feasible due to overfitting which can come with having too complex of a model design. You might want to try using a DT as a stacker, this is often a popular choice.
7