The Altair Community is migrating to a new platform to provide a better experience for you. In preparation for the migration, the Altair Community is on read-only mode from October 28 - November 6, 2024. Technical support via cases will continue to work as is. For any urgent requests from Students/Faculty members, please submit the form linked here
ModelApplier on multiple Models
Legacy User
Member Posts: 0 Newbie
Hi,
I built up a DecisionTree-Model on a Training-Dataset. The Validation is done by a XValidation. After writing down the model I run it over the Test-Dataset with the ModelApplier. This whole Processchain runs perfectly. But: My idea is to find the model which perfectly fits to the Test-Dataset. So I like to build up multiple models by implementing Bagging and evaluate them with the Testdata. The problem is that the ModelApplier can only handle one Model. Do you see an option to run multiple models on a Test-Dataset and evaluate them by ClassificationPerformance?
Regards,
Thorsten
I built up a DecisionTree-Model on a Training-Dataset. The Validation is done by a XValidation. After writing down the model I run it over the Test-Dataset with the ModelApplier. This whole Processchain runs perfectly. But: My idea is to find the model which perfectly fits to the Test-Dataset. So I like to build up multiple models by implementing Bagging and evaluate them with the Testdata. The problem is that the ModelApplier can only handle one Model. Do you see an option to run multiple models on a Test-Dataset and evaluate them by ClassificationPerformance?
Regards,
Thorsten
0
Answers
you could use the %{a}-macro together with an IteratingOperatorChain like it is described in this posting:
http://rapid-i.com/rapidforum/index.php/topic,32.0.html
This should also work in combination with a ClassificationPerformance evaluator - at least to manually check which model is the best one. For an completely automated selection, this would need a little amount of coding...
By the way: are you sure that it is a good idea to select the best model on the test set? This is actually like overfitting but now not on the training but on the test set. In general, I would always suggest to use all data for model building and use a validation scheme like cross validation for performance evaluation only but not for model selection...
Cheers,
Ingo
I also used the hole dataset for calculating a model. This is surely the best solution for dicriminating all the classes for this dataset. But in my case the model gets quite too complex because of high variations within the classes. So I would like to have an easier one which could also be used for similiar data.
Thanks so far,
Thorsten
Cheers,
Ingo