The company is developing an ML model to predict customer churn, a binary classification task (churn or no churn). The F1 score is an evaluation metric that balances precision and recall, making it suitable for assessing the performance of binary classification models, especially when dealing with imbalanced datasets, which is common in churn prediction.
Exact Extract from AWS AI Documents:
From the Amazon SageMaker Developer Guide:
"The F1 score is a metric for evaluating binary classification models, combining precision and recall into a single value. It is particularly useful for tasks like churn prediction, where class imbalance may exist, ensuring the model performs well on both positive and negative classes."
(Source: Amazon SageMaker Developer Guide, Model Evaluation Metrics)
Detailed Explanation:
Option A: F1 scoreThis is the correct answer. The F1 score is ideal for binary classification tasks like churn prediction, as it measures the model’s ability to correctly identify both churners and non-churners.
Option B: Mean squared error (MSE)MSE is used for regression tasks to measure the average squared difference between predicted and actual values, not for binary classification.
Option C: R-squaredR-squared is a metric for regression models, indicating how well the model explains the variability of the target variable. It is not applicable to classification tasks.
Option D: Time used to train the modelTraining time is not an evaluation metric for model performance; it measures the duration of training, not the model’s accuracy or effectiveness.
[References:, Amazon SageMaker Developer Guide: Model Evaluation Metrics (https://docs.aws.amazon.com/sagemaker/latest/dg/model-evaluation.html), AWS AI Practitioner Learning Path: Module on Model Performance and Evaluation, AWS Documentation: Metrics for Classification (https://aws.amazon.com/machine-learning/), , , , ]