Hey,
I have a question concerning the way you calculate metrics:
Predictions of trained model can vary from close to 0 to over 60 for some experiments that I performed using Kaggle Credit Card Fraud Dataset.
My question is: why do use the predictions as if they were probabilities (even though they are not withing [0, 1] range) when calculating AUC-ROC
and AUC-PR
?
I know that both sklearn
functions support this type of mixed input (binary and continuous vectors) but doesn't it give a false result?
This is a bit similar to what is happening in your implementation. Also, I think that the confidence threshold you mentioned in paper should be dependent on the value of margin used in deviation loss - not only on probit and normal distribution parameters.
Thanks for the reproducible paper :)
Kuba