# AI-900: AUC Score Explained: How to Interpret Binary Classification Models

Learn what AUC score means and how to use it to evaluate binary classification models. Find out how AUC score relates to ROC curve, TPR, FPR, accuracy, and variance.

## Question

You use an Azure Machine Learning designer pipeline to train and test a binary classification model. You review the model’s performance metrics in an Evaluate Model module, and note that it has an AUC score of 0.6. What can you conclude about the model?

A. The model performs better than random guessing
B. The model predicts accurately for 40% of cases
C. The model can explain 60% of the variance between true and predicted labels.

A. The model performs better than random guessing

## Explanation

The higher the score of AUC, the better the performance of the model.

The correct answer is A. The model performs better than random guessing.

The AUC score is a measure of how well a binary classification model can distinguish between the positive and negative classes. It is calculated by plotting the True Positive Rate (TPR) against the False Positive Rate (FPR) for all possible threshold values, and finding the area under the ROC curve. The ROC curve is a graph that shows the trade-off between sensitivity (TPR) and specificity (1-FPR) for a binary classifier.

The AUC score ranges from 0 to 1, where 0 means the model is completely wrong and 1 means the model is completely right. A random guessing model would have an AUC score of 0.5, meaning it has no ability to discriminate between the classes. Therefore, a model with an AUC score of 0.6 performs better than random guessing, but not by much. It means that the model has a 60% chance of ranking a random positive example higher than a random negative example.

The AUC score does not tell us anything about the accuracy of the model, which is the percentage of correct predictions out of all predictions. The accuracy depends on the chosen threshold value, which determines how the model classifies the examples based on the predicted probabilities or scores. The AUC score also does not tell us anything about the variance or the error of the model, which are measures of how much the model deviates from the true labels or the expected value.

Microsoft Azure AI Fundamentals AI-900 certification exam practice question and answer (Q&A) dump with detail explanation and reference available free, helpful to pass the Microsoft Azure AI Fundamentals AI-900 exam and earn Microsoft Azure AI Fundamentals AI-900 certification.

### Alex Lim

Alex Lim is a certified IT Technical Support Architect with over 15 years of experience in designing, implementing, and troubleshooting complex IT systems and networks. He has worked for leading IT companies, such as Microsoft, IBM, and Cisco, providing technical support and solutions to clients across various industries and sectors. Alex has a bachelor’s degree in computer science from the National University of Singapore and a master’s degree in information security from the Massachusetts Institute of Technology. He is also the author of several best-selling books on IT technical support, such as The IT Technical Support Handbook and Troubleshooting IT Systems and Networks. Alex lives in Bandar, Johore, Malaysia with his wife and two chilrdren. You can reach him at [email protected] or follow him on Website | Twitter | Facebook