The Binary Classification Evaluation component is used to calculate AUC, KS, and F1 score metrics to generate Kolmogorov–Smirnov (KS) curves, precision-recall (P-R) curves, ROC curves, lift charts, and gain charts.

Configure the component

You can use one of the following methods to configure the Binary Classification Evaluation component.

Method 1: Configure the component on the pipeline page

You can configure the parameters of the Binary Classification Evaluation component on the pipeline page of Machine Learning Designer of Machine Learning Platform for AI (PAI). Machine Learning Designer is formerly known as Machine Learning Studio. The following table describes the parameters.
ParameterDescription
Original Label ColumnThe name of the objective column.
Score ColumnThe prediction score column. Default value: prediction_score.
Positive Sample LabelSpecifies whether the samples are positive samples.
Number of Bins with Same Frequency when Calculating Indexes such as KS and PRThe number of bins obtained by using the equal frequency binning method.
Grouping ColumnThe group ID column. This parameter is used to calculate evaluation metrics for each group.
Advanced OptionsIf you select Advanced Options, the Prediction Result Detail Column, Prediction Targets Consistent With Evaluation Targets, and Save Performance Index parameters are valid.
Prediction Result Detail ColumnThe name of the prediction result detail column.
Prediction Targets Consistent with Evaluation TargetsSpecifies whether the prediction objective is consistent with the evaluation objective. For example, in a financial scenario, a program is trained to predict the probability that a customer is bad. The larger the probability is, the more likely the customer is bad. Related metrics such as lift evaluate the bad-customer detection rate. In this case, the prediction objective is consistent with the evaluation objective. In a credit scoring scenario, a program is trained to predict the probability that a customer is good. The larger the probability is, the more likely the customer is good. However, related metrics evaluate the bad-customer detection rate. In this case, the prediction objective is inconsistent with the evaluation objective.
Save Performance IndexSpecifies whether to save performance metrics.

Method 2: Use PAI commands

Configure the component parameters by using PAI commands. You can use the SQL Script component to call PAI commands. For more information, see SQL Script.
PAI -name=evaluate -project=algo_public
    -DoutputMetricTableName=output_metric_table
    -DoutputDetailTableName=output_detail_table
    -DinputTableName=input_data_table
    -DlabelColName=label
    -DscoreColName=score
ParameterRequiredDescriptionDefault value
inputTableNameYesThe name of the input table. N/A
inputTablePartitionsNoThe partitions that are selected from the input table for training. Full table
labelColNameYesThe name of the objective column. N/A
scoreColNameYesThe name of the score column. N/A
groupColNameNoThe name of the group column. This parameter is used for the evaluation of each group. N/A
binCountNoThe number of bins obtained by using the equal frequency binning method during the calculation of metrics such as KS and PR. 1000
outputMetricTableNameYesThe output metric table. The metrics include AUC, KS, and F1 score. N/A
outputDetailTableNameNoThe detail data table that is generated. N/A
positiveLabelNoSpecifies whether the samples are positive samples. 1
lifecycleNoThe lifecycle of the output table. N/A
coreNumNoThe number of cores. Determined by the system
memSizePerCoreNoThe memory size of each core. Determined by the system