
|
|||||||||||||||||||||||||||
AWS Certified AI Practitioner
AWS Certified AI Practitioner validates in-demand knowledge of artificial
intelligence (AI), machine learning (ML), and generative AI concepts and use
cases. Sharpen your competitive edge and position yourself for career growth and
higher earnings.
The AIF-C01 exam, AWS Certified AI Practitioner, is a foundational certification
that costs $100 USD and has a 90-minute duration with 65 questions. It is
designed for individuals with a general understanding of AI/ML and generative AI
on AWS, and the exam format includes multiple choice, multiple response,
ordering, matching, and case study questions. The exam is available in English
and other languages and can be taken at a Pearson VUE testing center or online.
Exam details
Exam code: AIF-C01
Exam name: AWS Certified AI Practitioner
Cost: $100 USD
Duration: 90 minutes
Total questions: 65 (50 scored, 15 unscored)
Question types: Multiple choice, multiple response, ordering, matching, and case
study
Passing score: 700 on a scaled score of 100–1,000
Mode: Available at Pearson VUE testing centers or online
Languages: English, Japanese, Korean, Portuguese (Brazil), and Simplified
Chinese
The exam has the following content domains and weightings:
• Domain 1: Fundamentals of AI and ML (20% of scored content)
• Domain 2: Fundamentals of Generative AI (24% of scored content)
• Domain 3: Applications of Foundation Models (28% of scored content)
• Domain 4: Guidelines for Responsible AI (14% of scored content)
• Domain 5: Security, Compliance, and Governance for AI Solutions (14% of scored content)
AIF-C01 Brain Dumps Exam + Online / Offline and Android Testing Engine & 4500+ other exams included
$50 - $25 (you save $25)
Buy Now
QUESTION 1
An AI practitioner trained a custom model on Amazon Bedrock by using a
training dataset that
contains confidential dat a. The AI practitioner wants to ensure that the custom
model does not generate inference responses
based on confidential data.
How should the AI practitioner prevent responses based on confidential data?
A. Delete the custom model. Remove the confidential data from the training
dataset. Retrain the custom model.
B. Mask the confidential data in the inference responses by using dynamic data
masking.
C. Encrypt the confidential data in the inference responses by using Amazon
SageMaker.
D. Encrypt the confidential data in the custom model by using AWS Key Management
Service (AWS KMS).
Answer: A
Explanation:
When a model is trained on a dataset containing confidential or sensitive data,
the model may
inadvertently learn patterns from this data, which could then be reflected in
its inference responses.
To ensure that a model does not generate responses based on confidential data,
the most effective
approach is to remove the confidential data from the training dataset and then
retrain the model.
Explanation of Each Option:
Option A (Correct): "Delete the custom model. Remove the confidential data from
the training
dataset. Retrain the custom model."This option is correct because it directly
addresses the core issue:
the model has been trained on confidential data. The only way to ensure that the
model does not
produce inferences based on this data is to remove the confidential information
from the training
dataset and then retrain the model from scratch. Simply deleting the model and
retraining it ensures
that no confidential data is learned or retained by the model. This approach
follows the best
practices recommended by AWS for handling sensitive data when using machine
learning services
like Amazon Bedrock.
Option B: "Mask the confidential data in the inference responses by using
dynamic data
masking."This option is incorrect because dynamic data masking is typically used
to mask or
obfuscate sensitive data in a database. It does not address the core problem of
the model
beingtrained on confidential data. Masking data in inference responses does not
prevent the model
from using confidential data it learned during training.
Option C: "Encrypt the confidential data in the inference responses by using
Amazon
SageMaker."This option is incorrect because encrypting the inference responses
does not prevent the
model from generating outputs based on confidential data. Encryption only
secures the data at rest
or in transit but does not affect the model's underlying knowledge or training
process.
Option D: "Encrypt the confidential data in the custom model by using AWS Key
Management Service
(AWS KMS)."This option is incorrect as well because encrypting the data within
the model does not
prevent the model from generating responses based on the confidential data it
learned during
training. AWS KMS can encrypt data, but it does not modify the learning that the
model has already performed.
AWS AI Practitioner Reference:
Data Handling Best Practices in AWS Machine Learning: AWS advises practitioners
to carefully handle
training data, especially when it involves sensitive or confidential
information. This includes
preprocessing steps like data anonymization or removal of sensitive data before
using it to train
machine learning models.
Amazon Bedrock and Model Training Security: Amazon Bedrock provides foundational
models and
customization capabilities, but any training involving sensitive data should
follow best practices, such
as removing or anonymizing confidential data to prevent unintended data leakage.
QUESTION 2
Which feature of Amazon OpenSearch Service gives companies the ability to build
vector database applications?
A. Integration with Amazon S3 for object storage
B. Support for geospatial indexing and queries
C. Scalable index management and nearest neighbor search capability
D. Ability to perform real-time analysis on streaming data
Answer: C
Explanation:
Amazon OpenSearch Service (formerly Amazon Elasticsearch Service) has introduced
capabilities to
support vector search, which allows companies to build vector database
applications. This is
particularly useful in machine learning, where vector representations
(embeddings) of data are often
used to capture semantic meaning.
Scalable index management and nearest neighbor search capability are the core
features enabling
vector database functionalities in OpenSearch. The service allows users to index
high-dimensional
vectors and perform efficient nearest neighbor searches, which are crucial for
tasks such as
recommendation systems, anomaly detection, and semantic search.
Here is why option C is the correct answer:
Scalable Index Management: OpenSearch Service supports scalable indexing of
vector data. This
means you can index a large volume of high-dimensional vectors and manage these
indexes in a costeffective
and performance-optimized way. The service leverages underlying AWS
infrastructure to
ensure that indexing scales seamlessly with data size.
Nearest Neighbor Search Capability: OpenSearch Service's nearest neighbor search
capability allows
for fast and efficient searches over vector data. This is essential for
applications like product
recommendation engines, where the system needs to quickly find the most similar
items based on a
user's query or behavior.
AWS AI Practitioner Reference:
According to AWS documentation, OpenSearch Service's support for nearest
neighbor search using
vector embeddings is a key feature for companies building machine learning
applications that
require similarity search.
The service uses Approximate Nearest Neighbors (ANN) algorithms to speed up
searches over large
datasets, ensuring high performance even with large-scale vector data.
The other options do not directly relate to building vector database
applications:
A . Integration with Amazon S3 for object storage is about storing data objects,
not vector-based
searching or indexing.
B . Support for geospatial indexing and queries is related to location-based
data, not vectors used in
machine learning.
D . Ability to perform real-time analysis on streaming data relates to analyzing
incoming data
streams, which is different from the vector search capabilities.
QUESTION 3
A company wants to display the total sales for its top-selling products
across various retail locations in the past 12 months.
Which AWS solution should the company use to automate the generation of graphs?
A. Amazon Q in Amazon EC2
B. Amazon Q Developer
C. Amazon Q in Amazon QuickSight
D. Amazon Q in AWS Chatbot
Answer: C
Explanation:
Amazon QuickSight is a fully managed business intelligence (BI) service that
allows users to create
and publish interactive dashboards that include visualizations like graphs,
charts, and tables.
"Amazon Q" is the natural language query feature within Amazon QuickSight. It
enables users to ask
questions about their data in natural language and receive visual responses such
as graphs.
Option C (Correct): "Amazon Q in Amazon QuickSight": This is the correct answer
because Amazon
QuickSight Q is specifically designed to allow users to explore their data
through natural language
queries, and it can automatically generate graphs to display sales data and
other metrics. This makes
it an ideal choice for the company to automate the generation of graphs showing
total sales for its
top-selling products across various retail locations.
Option A, B, and D: These options are incorrect:
A . Amazon Q in Amazon EC2: Amazon EC2 is a compute service that provides
virtual servers, but it is
not directly related to generating graphs or providing natural language querying
features.
B . Amazon Q Developer: This is not an existing AWS service or feature.
D . Amazon Q in AWS Chatbot: AWS Chatbot is a service that integrates with
Amazon Chime and
Slack for monitoring and managing AWS resources, but it is not used for
generating graphs based on sales data.
AWS AI Practitioner Reference:
Amazon QuickSight Q is designed to provide insights from data by using natural
language queries,
making it a powerful tool for generating automated graphs and visualizations
directly from queried data.
Business Intelligence (BI) on AWS: AWS services such as Amazon QuickSight
provide business
intelligence capabilities, including automated reporting and visualization
features, which are ideal
for companies seeking to visualize data like sales trends over time.
QUESTION 4
A company wants to build an interactive application for children that
generates new stories based on
classic stories. The company wants to use Amazon Bedrock and needs to ensure
that the results and
topics are appropriate for children.
Which AWS service or feature will meet these requirements?
A. Amazon Rekognition
B. Amazon Bedrock playgrounds
C. Guardrails for Amazon Bedrock
D. Agents for Amazon Bedrock
Answer: C
Explanation:
Amazon Bedrock is a service that provides foundational models for building
generative AI
applications. When creating an application for children, it is crucial to ensure
that the generated
content is appropriate for the target audience. "Guardrails" in Amazon Bedrock
provide mechanisms
to control the outputs and topics of generated content to align with desired
safety standards and
appropriateness levels.
Option C (Correct): "Guardrails for Amazon Bedrock": This is the correct answer
because guardrails
are specifically designed to help users enforce content moderation, filtering,
and safety checks on
the outputs generated by models in Amazon Bedrock. For a childrens application,
guardrails ensure
that all content generated is suitable and appropriate for the intended
audience.
Option A: "Amazon Rekognition" is incorrect. Amazon Rekognition is an image and
video analysis
service that can detect inappropriate content in images or videos, but it does
not handle text or story generation.
Option B: "Amazon Bedrock playgrounds" is incorrect because playgrounds are
environments for
experimenting and testing model outputs, but they do not inherently provide
safeguards to ensure
content appropriateness for specific audiences, such as children.
Option D: "Agents for Amazon Bedrock" is incorrect. Agents in Amazon Bedrock
facilitate building AI
applications with more interactive capabilities, but they do not provide
specific guardrails for
ensuring content appropriateness for children.
AWS AI Practitioner Reference:
Guardrails in Amazon Bedrock: Designed to help implement controls that ensure
generated content
is safe and suitable for specific use cases or audiences, such as children, by
moderating and filtering
inappropriate or undesired content.
Building Safe AI Applications: AWS provides guidance on implementing ethical AI
practices, including
using guardrails to protect against generating inappropriate or biased content.
QUESTION 5
A company has developed an ML model for image classification. The company wants
to deploy the
model to production so that a web application can use the model.
The company needs to implement a solution to host the model and serve
predictions without
managing any of the underlying infrastructure.
Which solution will meet these requirements?
A. Use Amazon SageMaker Serverless Inference to deploy the model.
B. Use Amazon CloudFront to deploy the model.
C. Use Amazon API Gateway to host the model and serve predictions.
D. Use AWS Batch to host the model and serve predictions.
Answer: A
Explanation:
Amazon SageMaker Serverless Inference is the correct solution for deploying an
ML model to
production in a way that allows a web application to use the model without the
need to manage the
underlying infrastructure.
Amazon SageMaker Serverless Inference provides a fully managed environment for
deploying
machine learning models. It automatically provisions, scales, and manages the
infrastructure
required to host the model, removing the need for the company to manage servers
or other
underlying infrastructure.
Why Option A is Correct:
No Infrastructure Management: SageMaker Serverless Inference handles the
infrastructure
management for deploying and serving ML models. The company can simply provide
the model and