I have been working with precision-recall curves and have trouble understanding how thresholds are decided.
This is my code:
import sklearn
precision, recall, thresholds = sklearn.metrics.precision_recall_curve(y_test,
probas_pred[:,1], pos_label=1, sample_weight=None)
which yields
precision = array([ 0.99971396, 1. , 1. , 1. , 1. , 1. , 1. ])
recall = array([ 1. , 0.99885551, 0.99341917, 0.96852647, 0.88898426, 0.70872675, 0. ])
thresholds = array[ 0.5, 0.6, 0.7, 0.8, 0.9, 1. ])
If I do np.unique(probas_pred[:,1]) (Random forest, high class imbalance) I get the following thresholds:
thresholds_probas_pred = array([ 0., 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 1.])
I thought that the precision recall curve plots Precision and Recall for all unique values in the probas_pred array. In this case the threshold values returned by precision-recall curve seems to ignore values less than 0.5. Could someone please explain ?
Thanks!
shape = [n_thresholds <= len(np.unique(probas_pred))]which is not consistent with the number of unique probabilities you are reporting in your question. - MhFarahani