The Altair Community is migrating to a new platform to provide a better experience for you. In preparation for the migration, the Altair Community is on read-only mode from October 28 - November 6, 2024. Technical support via cases will continue to work as is. For any urgent requests from Students/Faculty members, please submit the form linked here
What are most important parameters to tune for Deep Learning, XGBT, and Gen. Linear Model?
hi,
I want to try out those 3 new algorithms that came with 7.2 on my dataset (4500 examples with 25 num. attributes), what are the most important parameters to tune in a grid optimization operator for them? and in what intervals? are there any experiences..?
Tagged:
1
Answers
No free hunch :smileywink:
For deel learning, it basically depend on the network design and specific domain knowledge,
how to choose activation function, # epochs, hidden layer sizes, learning rate, parameters for avoid overfitting etc....
Why not download the booklet and take a look at the reference for the supervised models you just mentioned from
http://www.h2o.ai/docs/
you will get more helpful information there
Yea I'd like to get some ideas about the best params and their ranges to start tweaking with.
I run some sweeps and was rather disappointed.
I cannot run 10 params so any pointers are welcome.
Also surprisingly I get a better generalization of a smaller set than a bigger one (my total data set is just a few thousands of examples), what gives..?!