Amazon AIF-C01 AWS Certified AI Practitioner Exam Practice Test

Page: 1 / 14
Total 177 questions
Question 1

[AI and ML Concepts]

A financial institution is building an AI solution to make loan approval decisions by using a foundation model (FM). For security and audit purposes, the company needs the AI solution's decisions to be explainable.

Which factor relates to the explainability of the AI solution's decisions?



Answer : A

The financial institution needs an AI solution for loan approval decisions to be explainable for security and audit purposes. Explainability refers to the ability to understand and interpret how a model makes decisions. Model complexity directly impacts explainability: simpler models (e.g., logistic regression) are more interpretable, while complex models (e.g., deep neural networks) are harder to explain, often behaving like 'black boxes.'

Exact Extract from AWS AI Documents:

From the Amazon SageMaker Developer Guide:

'Model complexity affects the explainability of AI solutions. Simpler models, such as linear regression, are inherently more interpretable, while complex models, such as deep neural networks, may require additional tools like SageMaker Clarify to provide insights into their decision-making processes.'

(Source: Amazon SageMaker Developer Guide, Explainability with SageMaker Clarify)

Detailed

Option A: Model complexityThis is the correct answer. The complexity of the model directly influences how easily its decisions can be explained, a critical factor for audit and security purposes in loan approvals.

Option B: Training timeTraining time refers to how long it takes to train the model, which does not directly impact the explainability of its decisions.

Option C: Number of hyperparametersWhile hyperparameters affect model performance, they do not directly relate to explainability. A model with many hyperparameters might still be explainable if it is a simple model.

Option D: Deployment timeDeployment time refers to the time taken to deploy the model to production, which is unrelated to the explainability of its decisions.


Amazon SageMaker Developer Guide: Explainability with SageMaker Clarify (https://docs.aws.amazon.com/sagemaker/latest/dg/clarify-explainability.html)

AWS AI Practitioner Learning Path: Module on Responsible AI and Explainability

AWS Documentation: Explainable AI (https://aws.amazon.com/machine-learning/responsible-ai/)

Question 2

[AI and ML Concepts]

A company is developing an ML model to predict customer churn.

Which evaluation metric will assess the model's performance on a binary classification task such as predicting chum?



Answer : A

The company is developing an ML model to predict customer churn, a binary classification task (churn or no churn). The F1 score is an evaluation metric that balances precision and recall, making it suitable for assessing the performance of binary classification models, especially when dealing with imbalanced datasets, which is common in churn prediction.

Exact Extract from AWS AI Documents:

From the Amazon SageMaker Developer Guide:

'The F1 score is a metric for evaluating binary classification models, combining precision and recall into a single value. It is particularly useful for tasks like churn prediction, where class imbalance may exist, ensuring the model performs well on both positive and negative classes.'

(Source: Amazon SageMaker Developer Guide, Model Evaluation Metrics)

Detailed

Option A: F1 scoreThis is the correct answer. The F1 score is ideal for binary classification tasks like churn prediction, as it measures the model's ability to correctly identify both churners and non-churners.

Option B: Mean squared error (MSE)MSE is used for regression tasks to measure the average squared difference between predicted and actual values, not for binary classification.

Option C: R-squaredR-squared is a metric for regression models, indicating how well the model explains the variability of the target variable. It is not applicable to classification tasks.

Option D: Time used to train the modelTraining time is not an evaluation metric for model performance; it measures the duration of training, not the model's accuracy or effectiveness.


Amazon SageMaker Developer Guide: Model Evaluation Metrics (https://docs.aws.amazon.com/sagemaker/latest/dg/model-evaluation.html)

AWS AI Practitioner Learning Path: Module on Model Performance and Evaluation

AWS Documentation: Metrics for Classification (https://aws.amazon.com/machine-learning/)

Question 3

[AI and ML Concepts]

A company trained an ML model on Amazon SageMaker to predict customer credit risk. The model shows 90% recall on training data and 40% recall on unseen testing data.

Which conclusion can the company draw from these results?



Answer : A

The ML model shows 90% recall on training data but only 40% recall on unseen testing data, indicating a significant performance drop. This discrepancy suggests the model has learned the training data too well, including noise and specific patterns that do not generalize to new data, which is a classic sign of overfitting.

Exact Extract from AWS AI Documents:

From the Amazon SageMaker Developer Guide:

'Overfitting occurs when a model performs well on training data but poorly on unseen test data, as it has learned patterns specific to the training set, including noise, that do not generalize. A large gap between training and testing performance metrics, such as recall, is a common indicator of overfitting.'

(Source: Amazon SageMaker Developer Guide, Model Evaluation and Overfitting)

Detailed

Option A: The model is overfitting on the training data.This is the correct answer. The significant drop in recall from 90% (training) to 40% (testing) indicates the model is overfitting, as it performs well on training data but fails to generalize to unseen data.

Option B: The model is underfitting on the training data.Underfitting occurs when the model performs poorly on both training and testing data due to insufficient learning. With 90% recall on training data, the model is not underfitting.

Option C: The model has insufficient training data.Insufficient training data could lead to poor performance, but the high recall on trainingdata (90%) suggests the model has learned the training data well, pointing to overfitting rather than a lack of data.

Option D: The model has insufficient testing data.Insufficient testing data might lead to unreliable test metrics, but it does not explain the large performance gap between training and testing, which is more indicative of overfitting.


Amazon SageMaker Developer Guide: Model Evaluation and Overfitting (https://docs.aws.amazon.com/sagemaker/latest/dg/model-evaluation.html)

AWS AI Practitioner Learning Path: Module on Model Performance and Evaluation

AWS Documentation: Understanding Overfitting and Underfitting (https://aws.amazon.com/machine-learning/)

Question 4

[AI and ML Concepts]

A social media company wants to use a large language model (LLM) for content moderation. The company wants to evaluate the LLM outputs for bias and potential discrimination against specific groups or individuals.

Which data source should the company use to evaluate the LLM outputs with the LEAST administrative effort?



Answer : D

Benchmark datasets are pre-validated datasets specifically designed to evaluate machine learning models for bias, fairness, and potential discrimination. These datasets are the most efficient tool for assessing an LLM's performance against known standards with minimal administrative effort.

Option D (Correct): 'Benchmark datasets': This is the correct answer because using standardized benchmark datasets allows the company to evaluate model outputs for bias with minimal administrative overhead.

Option A: 'User-generated content' is incorrect because it is unstructured and would require significant effort to analyze for bias.

Option B: 'Moderation logs' is incorrect because they represent historical data and do not provide a standardized basis for evaluating bias.

Option C: 'Content moderation guidelines' is incorrect because they provide qualitative criteria rather than a quantitative basis for evaluation.

AWS AI Practitioner Reference:

Evaluating AI Models for Bias on AWS: AWS supports using benchmark datasets to assess model fairness and detect potential bias efficiently.


Question 5

[AI and ML Concepts]

A company is building an ML model. The company collected new data and analyzed the data by creating a correlation matrix, calculating statistics, and visualizing the data.

Which stage of the ML pipeline is the company currently in?



Answer : C

Exploratory data analysis (EDA) involves understanding the data by visualizing it, calculating statistics, and creating correlation matrices. This stage helps identify patterns, relationships, and anomalies in the data, which can guide further steps in the ML pipeline.

Option C (Correct): 'Exploratory data analysis': This is the correct answer as the tasks described (correlation matrix, calculating statistics, visualizing data) are all part of the EDA process.

Option A: 'Data pre-processing' is incorrect because it involves cleaning and transforming data, not initial analysis.

Option B: 'Feature engineering' is incorrect because it involves creating new features from raw data, not analyzing the data's existing structure.

Option D: 'Hyperparameter tuning' is incorrect because it refers to optimizing model parameters, not analyzing the data.

AWS AI Practitioner Reference:

Stages of the Machine Learning Pipeline: AWS outlines EDA as the initial phase of understanding and exploring data before moving to more specific preprocessing, feature engineering, and model training stages.


Question 6

A company needs to use Amazon SageMaker AI for model training and inference. The company must comply with regulatory requirements to run SageMaker jobs in an isolated environment without internet access.

Which solution will meet these requirements?



Answer : B


Question 7

[AI and ML Concepts]

A company is using Amazon SageMaker Studio notebooks to build and train ML models. The company stores the data in an Amazon S3 bucket. The company needs to manage the flow of data from Amazon S3 to SageMaker Studio notebooks.

Which solution will meet this requirement?



Answer : C

To manage the flow of data from Amazon S3 to SageMaker Studio notebooks securely, using a VPC with an S3 endpoint is the best solution.

Amazon SageMaker and S3 Integration:

Configuring SageMaker to use a Virtual Private Cloud (VPC) with an S3 endpoint allows the data flow between Amazon S3 and SageMaker Studio notebooks to occur over a private network.

This setup ensures that traffic between SageMaker and S3 does not traverse the public internet, enhancing security and performance.

Why Option C is Correct:

Secure Data Transfer: Ensures secure, private connectivity between SageMaker and S3, reducing exposure to potential security risks.

Direct Access to S3: Using an S3 endpoint in a VPC allows direct access to data in S3 without leaving the AWS network.

Why Other Options are Incorrect:

A . Amazon Inspector: Focuses on identifying security vulnerabilities, not managing data flow.

B . Amazon Macie: Monitors for sensitive data but does not manage data flow between S3 and SageMaker.

D . S3 Glacier Deep Archive: Is a storage class for archiving data, not for managing active data flow.


Page:    1 / 14   
Total 177 questions