Black Friday Sale Special - Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: xmaspas7

Easiest Solution 2 Pass Your Certification Exams

AIF-C01 Amazon Web Services AWS Certified AI Practitioner Exam Free Practice Exam Questions (2025 Updated)

Prepare effectively for your Amazon Web Services AIF-C01 AWS Certified AI Practitioner Exam certification with our extensive collection of free, high-quality practice questions. Each question is designed to mirror the actual exam format and objectives, complete with comprehensive answers and detailed explanations. Our materials are regularly updated for 2025, ensuring you have the most current resources to build confidence and succeed on your first attempt.

Page: 3 / 4
Total 289 questions

A company is developing an ML model to make loan approvals. The company must implement a solution to detect bias in the model. The company must also be able to explain the model's predictions.

Which solution will meet these requirements?

A.

Amazon SageMaker Clarify

B.

Amazon SageMaker Data Wrangler

C.

Amazon SageMaker Model Cards

D.

AWS AI Service Cards

A company built a deep learning model for object detection and deployed the model to production.

Which AI process occurs when the model analyzes a new image to identify objects?

A.

Training

B.

Inference

C.

Model deployment

D.

Bias correction

An ecommerce company wants to improve search engine recommendations by customizing the results for each user of the company's ecommerce platform. Which AWS service meets these requirements?

A.

Amazon Personalize

B.

Amazon Kendra

C.

Amazon Rekognition

D.

Amazon Transcribe

A company is using a large language model (LLM) on Amazon Bedrock to build a chatbot. The chatbot processes customer support requests. To resolve a request, the customer and the chatbot must interact a few times.

Which solution gives the LLM the ability to use content from previous customer messages?

A.

Turn on model invocation logging to collect messages.

B.

Add messages to the model prompt.

C.

Use Amazon Personalize to save conversation history.

D.

Use Provisioned Throughput for the LLM.

A company wants to use a large language model (LLM) on Amazon Bedrock for sentiment analysis. The company needs the LLM to produce more consistent responses to the same input prompt.

Which adjustment to an inference parameter should the company make to meet these requirements?

A.

Decrease the temperature value

B.

Increase the temperature value

C.

Decrease the length of output tokens

D.

Increase the maximum generation length

A company wants to assess internet quality in remote areas of the world. The company needs to collect internet speed data and store the data in Amazon RDS. The company will analyze internet speed variation throughout each day. The company wants to create an AI model to predict potential internet disruptions.

Which type of data should the company collect for this task?

A.

Tabular data

B.

Text data

C.

Time series data

D.

Audio data

Which technique breaks a complex task into smaller subtasks that are sent sequentially to a large language model (LLM)?

A.

One-shot prompting

B.

Prompt chaining

C.

Tree of thoughts

D.

Retrieval Augmented Generation (RAG)

An AI company periodically evaluates its systems and processes with the help of independent software vendors (ISVs). The company needs to receive email message notifications when an ISV's compliance reports become available.

Which AWS service can the company use to meet this requirement?

A.

AWS Audit Manager

B.

AWS Artifact

C.

AWS Trusted Advisor

D.

AWS Data Exchange

A company stores millions of PDF documents in an Amazon S3 bucket. The company needs to extract the text from the PDFs, generate summaries of the text, and index the summaries for fast searching.

Which combination of AWS services will meet these requirements? (Select TWO.)

A.

Amazon Translate

B.

Amazon Bedrock

C.

Amazon Transcribe

D.

Amazon Polly

E.

Amazon Textract

An AI practitioner trained a custom model on Amazon Bedrock by using a training dataset that contains confidential data. The AI practitioner wants to ensure that the custom model does not generate inference responses based on confidential data.

How should the AI practitioner prevent responses based on confidential data?

A.

Delete the custom model. Remove the confidential data from the training dataset. Retrain the custom model.

B.

Mask the confidential data in the inference responses by using dynamic data masking.

C.

Encrypt the confidential data in the inference responses by using Amazon SageMaker.

D.

Encrypt the confidential data in the custom model by using AWS Key Management Service (AWS KMS).

HOTSPOT

Select the correct AI term from the following list for each statement. Each AI term should be selected one time. (Select THREE.)

• AI

• Deep learning

• ML

An online learning company with large volumes of education materials wants to use enterprise search.

A.

Amazon Comprehend

B.

Amazon Textract

C.

Amazon Kendra

D.

Amazon Personalize

Sated and order the steps from the following bat to correctly describe the ML Lifecycle for a new custom modal Select each step one time. (Select and order FOUR.)

• Define the business objective.

• Deploy the modal.

• Develop and tram the model.

• Process the data.

An AI practitioner must fine-tune an open source large language model (LLM) for text categorization. The dataset is already prepared.

Which solution will meet these requirements with the LEAST operational effort?

A.

Create a custom model training job in PartyRock on Amazon Bedrock.

B.

Use Amazon SageMaker JumpStart to create a training job.

C.

Use a custom script to run an Amazon SageMaker AI model training job.

D.

Create a Jupyter notebook on an Amazon EC2 instance. Use the notebook to train the model.

A digital devices company wants to predict customer demand for memory hardware. The company does not have coding experience or knowledge of ML algorithms and needs to develop a data-driven predictive model. The company needs to perform analysis on internal data and external data.

Which solution will meet these requirements?

A.

Store the data in Amazon S3. Create ML models and demand forecast predictions by using Amazon SageMaker built-in algorithms that use the data from Amazon S3.

B.

Import the data into Amazon SageMaker Data Wrangler. Create ML models and demand forecast predictions by using SageMaker built-in algorithms.

C.

Import the data into Amazon SageMaker Data Wrangler. Build ML models and demand forecast predictions by using an Amazon Personalize Trending-Now recipe.

D.

Import the data into Amazon SageMaker Canvas. Build ML models and demand forecast predictions by selecting the values in the data from SageMaker Canvas.

A company has built a solution by using generative AI. The solution uses large language models (LLMs) to translate training manuals from English into other languages. The company wants to evaluate the accuracy of the solution by examining the text generated for the manuals.

Which model evaluation strategy meets these requirements?

A.

Bilingual Evaluation Understudy (BLEU)

B.

Root mean squared error (RMSE)

C.

Recall-Oriented Understudy for Gisting Evaluation (ROUGE)

D.

F1 score

A bank is building a chatbot to answer customer questions about opening a bank account. The chatbot will use public bank documents to generate responses. The company will use Amazon Bedrock and prompt engineering to improve the chatbot's responses.

Which prompt engineering technique meets these requirements?

A.

Complexity-based prompting

B.

Zero-shot prompting

C.

Few-shot prompting

D.

Directional stimulus prompting

A financial company uses a generative AI model to assign credit limits to new customers. The company wants to make the decision-making process of the model more transparent to its customers.

A.

Use a rule-based system instead of an ML model.

B.

Apply explainable AI techniques to show customers which factors influenced the model's decision.

C.

Develop an interactive UI for customers and provide clear technical explanations about the system.

D.

Increase the accuracy of the model to reduce the need for transparency.

A company is building a generative Al application and is reviewing foundation models (FMs). The company needs to consider multiple FM characteristics.

Select the correct FM characteristic from the following list for each definition. Each FM characteristic should be selected one time. (Select THREE.)

Concurrency

Context windows

Latency

What is tokenization used for in natural language processing (NLP)?

A.

To encrypt text data

B.

To compress text files

C.

To break text into smaller units for processing

D.

To translate text between languages

Page: 3 / 4
Total 289 questions
Copyright © 2014-2025 Solution2Pass. All Rights Reserved