According to the H2O documentation, the threshold used at prediction is the max F1 threshold from train. The performance function,
h2o.performance(model, newdata = test)
actually run the prediction on the test set in order to compute the confusion matrix.
Strangely I am getting different confusion matrix while predicting the same test set using :
h2o.predict(object, newdata=test).
It means that h2o.performance()
is using a different threshold from h2o.predict()
.
I am wondering how can i dictate the threshold upon prediction.