The Altair Community is migrating to a new platform to provide a better experience for you. In preparation for the migration, the Altair Community is on read-only mode from October 28 - November 6, 2024. Technical support via cases will continue to work as is. For any urgent requests from Students/Faculty members, please submit the form linked here
split results into bins for evaluation
Hi,
Another fun puzzle for the RM team.
I've run a model that outputs confidence values for an SVM class. The values range from 0 to 1 (as expected for this type of model.)
One very common method of evaluation I've seen in papers is to break the results into "bins" or "groups" by confidence range and then report the accuracy of each range.
Something like
Any way to this sort of analysis in RM??
Thanks!
Another fun puzzle for the RM team.
I've run a model that outputs confidence values for an SVM class. The values range from 0 to 1 (as expected for this type of model.)
One very common method of evaluation I've seen in papers is to break the results into "bins" or "groups" by confidence range and then report the accuracy of each range.
Something like
Range | # predicted | # correct | % correct |
60-65 | 35 | 21 | 60 |
55-60 | 130 | 75 | 57.6 |
Thanks!
Tagged:
0
Answers
I do not have the time to create a perfect fully tested process, but this non-executable process should give you an idea: I hope you can make it from here.
regards,
Steffen
Cheers,
Ingo
That's a great way to solve this problem.
I'm getting good predictions, but the confidence "score" seems out of alignment. For example, my 90%-100% confidence level is true about 70%. My 80-90% is true about 60%, etc. So, I guess the model is good, but the scores are just "scores" and not reliable probability.
Thanks!!!!!
-B
Nice to hear. As a quick idea, you could use the operator Platt Scaling to calibrate the scores i.e. to make them a better approximation to the true probabilties.
kind regards,
Steffen
A great idea as usual!
Question: Where in the process would I put the platt operator. Do I do it after training? or testing? or both?
Thnaks!
I suggest something like the process below. Additional remark: The XVPrediction is used to train a separate basis of confidences to prevent overfitting as suggested in platt's original paper. regards,
Steffen
VERY clever application.
I don't entirely understand why you train an SVM and then train 3 more svm with platt insie the xvpred?? Does the XVPrediciton somehow deliver the "best" platt from the XV tests?
Also, does the platt model from within the XVPrediction pass through to the model that we eventually apply? If so, would it be safe to assume that we could write that model and it would include the Platt as well?
hope this was helpful,
regards,
Steffen
PS: Sorry, the name is "Steffen", not "Stefan". This is an important difference here in Germany