TabularPredictor.evaluate_predictions#

TabularPredictor.evaluate_predictions(y_true, y_pred, sample_weight=None, silent=False, auxiliary_metrics=True, detailed_report=False) dict[source]#

Evaluate the provided prediction probabilities against ground truth labels. Evaluation is based on the eval_metric previously specified in init, or default metrics if none was specified.

Parameters
  • y_true (np.array or pd.Series) – The ordered collection of ground-truth labels.

  • y_pred (pd.Series or pd.DataFrame) – The ordered collection of prediction probabilities or predictions. Obtainable via the output of predictor.predict_proba. Caution: For certain types of eval_metric (such as ‘roc_auc’), y_pred must be predicted-probabilities rather than predicted labels.

  • sample_weight (pd.Series, default = None) – Sample weight for each row of data. If None, uniform sample weights are used.

  • silent (bool, default = False) – If False, performance results are printed.

  • auxiliary_metrics (bool, default = True) – Should we compute other (problem_type specific) metrics in addition to the default metric?

  • detailed_report (bool, default = False) – Should we computed more detailed versions of the auxiliary_metrics? (requires auxiliary_metrics = True)

Returns

  • Returns dict where keys = metrics, values = performance along each metric.

  • NOTE (Metrics scores always show in higher is better form.)

  • This means that metrics such as log_loss and root_mean_squared_error will have their signs FLIPPED, and values will be negative.