A company is developing a new model to predict the prices of specific items. The model performed well on the training dataset. When the company deployed the model to production, the model's performance decreased significantly.
What should the company do to mitigate this problem?
Answer : C
When a model performs well on the training data but poorly in production, it is often due to overfitting. Overfitting occurs when a model learns patterns and noise specific to the training data, which does not generalize well to new, unseen data in production. Increasing the volume of data used in training can help mitigate this problem by providing a more diverse and representative dataset, which helps the model generalize better.
Option C (Correct): 'Increase the volume of data that is used in training': Increasing the data volume can help the model learn more generalized patterns rather than specific features of the training dataset, reducing overfitting and improving performance in production.
Option A: 'Reduce the volume of data that is used in training' is incorrect, as reducing data volume would likely worsen the overfitting problem.
Option B: 'Add hyperparameters to the model' is incorrect because adding hyperparameters alone does not address the issue of data diversity or model generalization.
Option D: 'Increase the model training time' is incorrect because simply increasing training time does not prevent overfitting; the model needs more diverse data.
AWS AI Practitioner Reference:
Best Practices for Model Training on AWS: AWS recommends using a larger and more diverse training dataset to improve a model's generalization capability and reduce the risk of overfitting.
A hospital developed an AI system to provide personalized treatment recommendations for patients. The AI system must provide the rationale behind the recommendations and make the insights accessible to doctors and patients.
Which human-centered design principle does this scenario present?
Answer : A
Explainability refers to the ability of an AI system to make its decision-making process clear and understandable to humans.
A is correct:
'Explainability is crucial for human-centered AI, especially in healthcare, to ensure that doctors and patients understand the rationale behind AI-driven recommendations.'
(Reference: AWS Responsible AI)
B relates to protecting data, not explanations.
C is about treating groups equally.
D is about managing data lifecycle, not providing rationales.
A company trained an ML model on Amazon SageMaker to predict customer credit risk. The model shows 90% recall on training data and 40% recall on unseen testing data.
Which conclusion can the company draw from these results?
Answer : A
The ML model shows 90% recall on training data but only 40% recall on unseen testing data, indicating a significant performance drop. This discrepancy suggests the model has learned the training data too well, including noise and specific patterns that do not generalize to new data, which is a classic sign of overfitting.
Exact Extract from AWS AI Documents:
From the Amazon SageMaker Developer Guide:
'Overfitting occurs when a model performs well on training data but poorly on unseen test data, as it has learned patterns specific to the training set, including noise, that do not generalize. A large gap between training and testing performance metrics, such as recall, is a common indicator of overfitting.'
(Source: Amazon SageMaker Developer Guide, Model Evaluation and Overfitting)
Detailed
Option A: The model is overfitting on the training data.This is the correct answer. The significant drop in recall from 90% (training) to 40% (testing) indicates the model is overfitting, as it performs well on training data but fails to generalize to unseen data.
Option B: The model is underfitting on the training data.Underfitting occurs when the model performs poorly on both training and testing data due to insufficient learning. With 90% recall on training data, the model is not underfitting.
Option C: The model has insufficient training data.Insufficient training data could lead to poor performance, but the high recall on trainingdata (90%) suggests the model has learned the training data well, pointing to overfitting rather than a lack of data.
Option D: The model has insufficient testing data.Insufficient testing data might lead to unreliable test metrics, but it does not explain the large performance gap between training and testing, which is more indicative of overfitting.
Amazon SageMaker Developer Guide: Model Evaluation and Overfitting (https://docs.aws.amazon.com/sagemaker/latest/dg/model-evaluation.html)
AWS AI Practitioner Learning Path: Module on Model Performance and Evaluation
AWS Documentation: Understanding Overfitting and Underfitting (https://aws.amazon.com/machine-learning/)
A company wants to build an interactive application for children that generates new stories based on classic stories. The company wants to use Amazon Bedrock and needs to ensure that the results and topics are appropriate for children.
Which AWS service or feature will meet these requirements?
Answer : C
Amazon Bedrock is a service that provides foundational models for building generative AI applications. When creating an application for children, it is crucial to ensure that the generated content is appropriate for the target audience. 'Guardrails' in Amazon Bedrock provide mechanisms to control the outputs and topics of generated content to align with desired safety standards and appropriateness levels.
Option C (Correct): 'Guardrails for Amazon Bedrock': This is the correct answer because guardrails are specifically designed to help users enforce content moderation, filtering, and safety checks on the outputs generated by models in Amazon Bedrock. For a children's application, guardrails ensure that all content generated is suitable and appropriate for the intended audience.
Option A: 'Amazon Rekognition' is incorrect. Amazon Rekognition is an image and video analysis service that can detect inappropriate content in images or videos, but it does not handle text or story generation.
Option B: 'Amazon Bedrock playgrounds' is incorrect because playgrounds are environments for experimenting and testing model outputs, but they do not inherently provide safeguards to ensure content appropriateness for specific audiences, such as children.
Option D: 'Agents for Amazon Bedrock' is incorrect. Agents in Amazon Bedrock facilitate building AI applications with more interactive capabilities, but they do not provide specific guardrails for ensuring content appropriateness for children.
AWS AI Practitioner Reference:
Guardrails in Amazon Bedrock: Designed to help implement controls that ensure generated content is safe and suitable for specific use cases or audiences, such as children, by moderating and filtering inappropriate or undesired content.
Building Safe AI Applications: AWS provides guidance on implementing ethical AI practices, including using guardrails to protect against generating inappropriate or biased content.
A company wants to improve the accuracy of the responses from a generative AI application. The application uses a foundation model (FM) on Amazon Bedrock.
Which solution meets these requirements MOST cost-effectively?
Answer : D
The company wants to improve the accuracy of a generative AI application using a foundation model (FM) on Amazon Bedrock in the most cost-effective way. Prompt engineering involves optimizing the input prompts to guide the FM to produce more accurate responses without modifying the model itself. This approach is cost-effective because it does not require additional computational resources or training, unlike fine-tuning or retraining.
Exact Extract from AWS AI Documents:
From the AWS Bedrock User Guide:
'Prompt engineering is a cost-effective technique to improve the performance of foundation models. By crafting precise and context-rich prompts, users can guide the model to generate more accurate and relevant responses without the need for fine-tuning or retraining.'
(Source: AWS Bedrock User Guide, Prompt Engineering for Foundation Models)
Detailed
Option A: Fine-tune the FM.Fine-tuning involves retraining the FM on a custom dataset, which requirescomputational resources, time, and cost (e.g., for Amazon Bedrock fine-tuning jobs). It is not the most cost-effective solution.
Option B: Retrain the FM.Retraining an FM from scratch is highly resource-intensive and expensive, as it requires large datasets and significant compute power. This is not cost-effective.
Option C: Train a new FM.Training a new FM is the most expensive option, as it involves building a model from the ground up, requiring extensive data, compute resources, and expertise. This is not cost-effective.
Option D: Use prompt engineering.This is the correct answer. Prompt engineering adjusts the input prompts to improve the FM's responses without incurring additional compute costs, making it the most cost-effective solution for improving accuracy on Amazon Bedrock.
AWS Bedrock User Guide: Prompt Engineering for Foundation Models (https://docs.aws.amazon.com/bedrock/latest/userguide/prompt-engineering.html)
AWS AI Practitioner Learning Path: Module on Generative AI Optimization
Amazon Bedrock Developer Guide: Cost Optimization for Generative AI (https://aws.amazon.com/bedrock/)
Which AWS service or feature can help an AI development team quickly deploy and consume a foundation model (FM) within the team's VPC?
Answer : B
Amazon SageMaker JumpStart is the correct service for quickly deploying and consuming a foundation model (FM) within a team's VPC.
Amazon SageMaker JumpStart:
Provides access to a wide range of pre-trained models and solutions that can be easily deployed and consumed within a VPC.
Designed to simplify and accelerate the deployment of machine learning models, including foundation models.
Why Option B is Correct:
Rapid Deployment: JumpStart allows for quick deployment of models with minimal configuration, directly within a secure VPC environment.
Ease of Use: Provides a user-friendly interface to select and deploy models, reducing the time to value.
Why Other Options are Incorrect:
A . Amazon Personalize: Focuses on creating personalized recommendations, not deploying foundation models.
C . PartyRock: Not a recognized AWS service.
D . Amazon SageMaker endpoints: Endpoints are for deploying specific models, not a feature for quickly starting with pre-trained foundation models.
A company wants to use AI for budgeting. The company made one budget manually and one budget by using an AI model. The company compared the budgets to evaluate the performance of the AI model. The AI model budget produced incorrect numbers.
Which option represents the AI model's problem?
Answer : A
Comprehensive and Detailed Explanation From Exact AWS AI documents:
Hallucinations occur when an AI model generates incorrect, fabricated, or misleading outputs that appear plausible but are factually wrong.
AWS generative AI guidance identifies hallucinations as:
A common limitation of generative models
A risk when models generate numerical or factual data
A key reason for validation and human review in critical use cases
Why the other options are incorrect:
Safety (B) relates to harmful or restricted content.
Interpretability (C) refers to understanding how a model makes decisions.
Cost (D) concerns operational expenses.
AWS AI document references:
Generative AI Risks and Limitations
Responsible Use of Foundation Models