"roc_auc_score" is defined as the area under the ROC curve, which is the curve having False Positive Rate on the x-axis and True Positive Rate on the y-axis at all classification thresholds.
In other words - roc_auc_score coincides with “the probability that a classifier will rank a randomly chosen positive instance higher than a randomly chosen negative one”.
Classification-threshold invariance is not always desirable, in cases where there are wide disparities in the cost of false negatives vs. false positives, it may be critical to minimize one type of classification error.
e.g, in email spam detection, you likely want to prioritize minimizing false positives (i.e. an Email is NOT spam, but its positively determined as a spam and hence moved to span folder).
Even if that results in a significant increase of false negatives. (An email is indeed spam, but model determines it to be negative, i.e. Not-Spam). AUC isn't a useful metric for this type of optimization.
“One vs Rest” is a method to evaluate multiclass models by comparing each class against all the others at the same time. Here we take one class and consider it as our “positive” class, while all the others (the rest) are considered as the “negative” class.
e.s. if you have three classes named X, Y, and Z, you will have one ROC for X classified against Y and Z, another ROC for Y classified against X and Z, and the third one of Z classified against Y and X.
3/ If the p-value from the test is less than some significance level (e.g. α = .05), then we can reject the null hypothesis and conclude that the time series is stationary.
2/ It is important to standardize variables before running Cluster Analysis. It is because cluster analysis techniques depend on the concept of measuring the distance between the different observations we're trying to cluster.