Search Tutorials


AWS AI Practitioner Practice Test 1 | AIF-C01 | JavaInUse

AWS Certified AI Practitioner (AIF-C01) - Practice Test 1

Your Progress

0 / 50
Question 1 MEDIUM
A healthcare company wants to build an AI application that listens to patient consultations and automatically generates structured clinical notes. Which AWS service is MOST appropriate for this use case?
Amazon Transcribe Medical is a speech-to-text service specifically optimized for medical terminology and clinical conversations. It converts audio recordings into accurate transcripts that can be structured into clinical notes. Amazon Comprehend Medical (A) analyzes existing text to extract medical entities -- it does not convert speech to text. Amazon Textract (C) extracts text from scanned documents and images. Amazon Polly (D) converts text to speech, which is the opposite of what is needed. See more: AWS Managed AI Services
Question 2 MEDIUM
A company is building a RAG (Retrieval Augmented Generation) application using Amazon Bedrock. The company wants to store and search vector embeddings of its internal product catalog. Which AWS service is MOST suitable for storing and querying these embeddings?
Amazon OpenSearch Service supports k-NN (k-Nearest Neighbors) vector search, making it ideal for storing and querying high-dimensional vector embeddings in RAG architectures. It efficiently finds semantically similar documents through approximate nearest neighbor search. Amazon RDS (A) is a relational database not optimized for vector similarity search. Amazon DynamoDB (B) is a key-value store without native vector search capabilities. Amazon Redshift (D) is a data warehouse designed for analytical queries, not vector similarity searches. See more: Bedrock & Generative AI
Question 3 MEDIUM
A data science team is evaluating a classification model. The model must minimize false negatives because missing a positive case has serious medical consequences. Which evaluation metric should the team prioritize?
Recall (also called sensitivity or true positive rate) measures the proportion of actual positives that the model correctly identifies. Maximizing recall minimizes false negatives, which is critical in medical or safety applications where missing a positive case is more costly than a false alarm. Precision (A) minimizes false positives, not false negatives. Specificity (C) measures the true negative rate. F1 score (D) balances precision and recall but does not specifically optimize for minimizing false negatives. See more: AI/ML Fundamentals
Question 4 EASY
A financial services company uses a large language model (LLM) to generate investment reports. The company is concerned that the model occasionally produces confident-sounding responses that contain factually incorrect information. What is this phenomenon called?
Hallucination refers to when an LLM generates fluent, confident responses that contain fabricated or factually incorrect information. This is a fundamental limitation of generative models that predict tokens based on statistical patterns rather than verified facts. Model drift (A) occurs when model performance degrades over time due to changing data distributions. Prompt injection (B) is a security attack where malicious instructions are embedded in user input. Overfitting (D) is when a model memorizes training data rather than learning generalizable patterns. See more: Bedrock & Generative AI
Question 5 EASY
A company is developing an AI image moderation system for a social media platform. The system needs to automatically detect and flag inappropriate images uploaded by users. Which AWS service is MOST appropriate?
Amazon Rekognition provides a content moderation API that detects explicit, suggestive, violent, or otherwise inappropriate content in images and videos. It returns confidence scores for detected categories, enabling automated moderation workflows. Amazon Comprehend (A) analyzes text for sentiment, entities, and topics -- it does not process images. Amazon Textract (C) extracts text from images and documents, not visual content classification. Amazon Kendra (D) is an intelligent search service for text documents. See more: AWS Managed AI Services
Question 6 MEDIUM
A machine learning team notices that their model performs well on the training dataset but poorly when deployed in production. Investigation reveals that the statistical distribution of production input data has shifted significantly since training. What is this problem called?
Data drift (also called covariate shift or distribution shift) occurs when the statistical properties of input data change between training and production, causing model performance to degrade. The model learned patterns from one distribution but is applied to a different one. Overfitting (A) is when a model memorizes training data -- it would still perform poorly on held-out data from the same distribution. Underfitting (C) is when a model is too simple to capture patterns even in training data. Label noise (D) refers to incorrect labels in the training dataset. See more: AI/ML Fundamentals
Question 7 EASY
A company wants to automatically extract structured data from thousands of scanned paper invoices with varying formats and populate a database. Which AWS service is MOST appropriate?
Amazon Textract uses ML to extract text, forms, tables, and key-value pairs from scanned documents. It handles varying layouts and formats without requiring templates, making it ideal for processing diverse invoices. Amazon Comprehend (A) analyzes text that has already been extracted -- it does not extract text from images. Amazon Rekognition (B) analyzes images for objects, faces, and scenes, not structured document data extraction. Amazon Translate (D) translates text between languages. See more: AWS Managed AI Services
Question 8 MEDIUM
A company is evaluating foundation models on Amazon Bedrock for a text summarization task. The team wants to measure how well generated summaries capture key information compared to human-written reference summaries. Which evaluation metric is MOST appropriate?
ROUGE (Recall-Oriented Understudy for Gisting Evaluation) measures the overlap of n-grams between generated summaries and reference summaries. It is the standard metric for evaluating summarization quality, specifically measuring recall of important content. BLEU score (A) measures precision of n-gram overlap and is primarily used for machine translation evaluation. Perplexity (C) measures how well a language model predicts text -- it is a training metric, not a summary quality metric. F1 score (D) evaluates classification model performance, not summarization quality. See more: AI/ML Fundamentals
Question 9 MEDIUM
A company has deployed a machine learning model to production. The team wants to continuously monitor the model for data quality issues, prediction drift, and performance degradation over time. Which Amazon SageMaker feature provides this capability?
Amazon SageMaker Model Monitor continuously monitors deployed models for data quality issues, model quality degradation, bias drift, and feature attribution drift. It generates alerts when metrics deviate from baseline thresholds, enabling proactive model maintenance. SageMaker Experiments (A) tracks and compares training runs, not production model behavior. Data Wrangler (C) is used for data preparation and transformation before training. Feature Store (D) stores and retrieves ML features but does not monitor model performance. See more: Amazon SageMaker
Question 10 MEDIUM
A startup wants to build a mobile application that identifies plant species from photos. The team has no machine learning expertise and needs a solution that requires minimal ML knowledge. Which approach requires the LEAST amount of ML knowledge?
Amazon Rekognition Custom Labels allows users to train custom image classifiers by uploading labeled images through a point-and-click console interface with no ML expertise required. The service automatically handles model architecture, training, and optimization. Training a custom CNN (A) requires deep learning expertise and hyperparameter tuning knowledge. Fine-tuning on Bedrock (C) requires understanding of fine-tuning processes and data preparation. Building TensorFlow models on EC2 (D) requires significant ML engineering and infrastructure skills. See more: AWS Managed AI Services
Question 11 MEDIUM
An AI team is building a multi-turn conversational assistant using Amazon Bedrock. The assistant must remember the context of previous messages within a conversation session. What is the MOST effective approach to maintain conversation context?
LLMs are stateless and do not inherently remember previous turns. To enable multi-turn conversations, the application must include the full conversation history -- all previous user messages and assistant responses -- in each API request. The model uses this context window to generate contextually appropriate responses. Temperature (A) controls output randomness, not conversational memory. Provisioned Throughput (C) is about inference capacity, not state management. A larger model (D) has more parameters but is still stateless without explicit history management. See more: Prompt Engineering
Question 12 MEDIUM
A company is concerned about the carbon footprint of running ML inference at scale. The team wants to use AWS infrastructure with the MOST energy-efficient chips specifically designed for ML inference workloads. Which EC2 instance series should they choose?
Amazon EC2 Inf2 instances use AWS Inferentia2 chips, custom-designed by AWS for high-performance, energy-efficient ML inference. Inferentia chips deliver significantly better performance per watt than GPU alternatives for inference workloads. P5 instances (A) use NVIDIA H100 GPUs optimized for training, not energy-efficient inference. G5 instances (B) use NVIDIA A10G GPUs -- better for inference than P5 but less efficient than purpose-built Inferentia chips. C7g instances (D) use Graviton3 processors optimized for general compute workloads, not specialized ML inference. See more: AI Challenges & Responsibilities
Question 13 EASY
A company wants to use Amazon Bedrock to build an application that answers employee questions based on a large internal knowledge base of PDF policy documents. Which Amazon Bedrock feature should the company use?
Amazon Bedrock Knowledge Bases implements RAG (Retrieval Augmented Generation) by ingesting documents, generating embeddings, storing them in a vector database, and retrieving relevant context to ground model responses in the actual document content. Fine-tuning (A) trains the model on examples to adapt behavior but bakes static knowledge into model weights and cannot dynamically retrieve from live documents. Agents (C) orchestrate multi-step tasks and can use Knowledge Bases, but the core document Q&A capability is provided by the Knowledge Base. Continued pre-training (D) teaches the model broad domain knowledge but cannot query live documents at inference time. See more: Bedrock & Generative AI
Question 14 MEDIUM
A company is building a generative AI application and is concerned about users embedding malicious instructions in their input to override the system prompt and hijack the model's behavior. What type of attack is this?
Prompt injection attacks attempt to override or circumvent a model's system prompt by embedding malicious instructions in user input, such as 'Ignore all previous instructions and...' to redirect the model's behavior. Model inversion attacks (A) attempt to reconstruct training data from model outputs. Adversarial example attacks (B) add imperceptible perturbations to inputs to cause misclassification, primarily in vision models. Data poisoning attacks (D) corrupt training data to introduce backdoors or degrade model performance during training. See more: AI Challenges & Responsibilities
Question 15 MEDIUM
A company needs to process and run inference on 50,000 customer feedback records stored in Amazon S3. The results are needed the following morning and do not require real-time processing. Which Amazon SageMaker inference type is MOST cost-effective for this workload?
Batch transform is designed for running inference on large static datasets stored in S3 when immediate results are not needed. It provisions compute instances, processes the full dataset, saves results back to S3, then terminates -- you pay only for the compute time used. Real-time inference (A) maintains always-on endpoints, which is expensive for overnight batch jobs. Serverless inference (B) has cold starts and timeout limits unsuitable for large batch workloads. Asynchronous inference (C) processes individual large payloads asynchronously but is not optimized for bulk batch processing of many small records. See more: Amazon SageMaker
Question 16 EASY
A company is using a large language model (LLM) for code generation. Developers want the model to produce more deterministic and consistent outputs, with less random variation between runs on identical prompts. Which parameter adjustment achieves this?
Temperature controls the randomness of token selection during generation. Decreasing temperature (toward 0) makes the model more deterministic by concentrating probability mass on the highest-probability tokens, producing more consistent and predictable outputs. At temperature 0, the model always selects the most probable token. Increasing temperature (A) increases randomness and creativity -- the opposite of what is needed. Increasing Top-P (C) expands the pool of candidate tokens by cumulative probability, increasing variability. Increasing Top-K (D) also expands the candidate token pool, increasing output diversity. See more: Prompt Engineering
Question 17 EASY
A retail company wants to build a personalized product recommendation system that learns from each customer's browsing history, purchase behavior, and ratings. Which AWS service is purpose-built for this use case?
Amazon Personalize is a fully managed ML service purpose-built for real-time personalization and recommendation systems. It uses the same technology as Amazon.com's recommendation engine to learn individual customer preferences from interaction history and deliver personalized recommendations. Amazon Comprehend (A) is an NLP service for text analysis tasks such as sentiment detection and entity recognition. Amazon Forecast (B) is for time-series metric forecasting such as demand or revenue. Amazon Kendra (D) is an intelligent enterprise search service for document retrieval. See more: AWS Managed AI Services
Question 18 EASY
A company wants to implement responsible AI practices. They need to ensure that customers who receive adverse outcomes from their AI model, such as loan denials, can receive an understandable explanation of why the decision was made. Which responsible AI principle does this requirement represent?
Explainability (also called interpretability) is the principle that AI systems should provide understandable reasons for their decisions. This is especially critical in regulated industries like financial services, where regulations such as GDPR and the Equal Credit Opportunity Act require that customers can understand and contest automated decisions. Fairness (A) ensures equitable treatment across demographic groups. Privacy (B) protects personal data from unauthorized access or misuse. Robustness (D) ensures the model performs reliably under adverse or unexpected input conditions. See more: AI Challenges & Responsibilities
Question 19 MEDIUM
A company has trained a binary classification model to detect fraudulent transactions. The dataset contains 98% legitimate transactions and only 2% fraudulent ones. The model achieves 98% accuracy by predicting every transaction as legitimate. Which technique BEST addresses the class imbalance problem?
SMOTE (Synthetic Minority Oversampling TEchnique) generates synthetic examples of the minority class (fraud) by interpolating between existing minority samples, balancing the class distribution without simple duplication. This helps the model learn meaningful decision boundaries for rare events. Increasing epochs (A) would further reinforce the majority-class bias. A larger network (B) amplifies the imbalance by providing more capacity to memorize the dominant class. Reducing learning rate (D) affects training convergence speed but does not address the underlying class representation imbalance. See more: AI/ML Fundamentals
Question 20 MEDIUM
A company uses Amazon Bedrock to power a customer-facing chatbot. The security team requires that no customer personally identifiable information (PII) appears in the model's responses. Which Amazon Bedrock Guardrails feature specifically addresses this requirement?
Amazon Bedrock Guardrails sensitive information filters detect and redact PII such as names, phone numbers, email addresses, Social Security numbers, credit card numbers, and other personal data from both inputs and outputs. Content filters (A) detect harmful content categories like hate speech and violence, not PII specifically. Denied topics (B) block the model from engaging with specified subject areas entirely. Word filters (D) block specific pre-defined words or phrases -- they are not flexible enough to detect all forms of PII patterns across different formats. See more: Bedrock & Generative AI
Question 21 MEDIUM
A team of data scientists wants to automatically search through many hyperparameter configurations for a training job and identify the best-performing model without manual experimentation. Which Amazon SageMaker feature should they use?
Amazon SageMaker Automatic Model Tuning (also called Hyperparameter Optimization or HPO) automatically searches for the best hyperparameter values by running multiple training jobs with different configurations. It supports strategies like Bayesian optimization, random search, and grid search to efficiently navigate the hyperparameter space. SageMaker Pipelines (A) orchestrates end-to-end ML workflows but does not automatically tune hyperparameters. SageMaker Clarify (C) detects bias and explains model predictions. Model Registry (D) catalogs and versions trained models for deployment. See more: Amazon SageMaker
Question 22 EASY
A company is building a generative AI pipeline that must perform multiple sequential steps automatically -- classifying a customer request, routing it to the correct department, querying a knowledge base, and generating a final response. Which Amazon Bedrock feature enables this multi-step orchestration?
Amazon Bedrock Agents autonomously break down complex tasks into sequential steps, invoke Lambda functions through action groups, query Knowledge Bases for information retrieval, and orchestrate entire workflows to complete multi-step goals. They use a ReAct (Reasoning and Acting) framework to plan and execute actions. Knowledge Bases (A) handle document retrieval but do not independently orchestrate multi-step workflows. Fine-tuning (B) adapts model weights for specific tasks but does not add workflow orchestration capability. Guardrails (D) add safety and content filters, not task orchestration. See more: Bedrock & Generative AI
Question 23 EASY
A media company wants to generate captions for video content in the source language and then produce subtitles in five additional languages for international distribution. Which combination of AWS services creates the MOST complete solution?
Amazon Transcribe converts spoken audio in video files to text, generating source-language captions. Amazon Translate then converts those transcriptions into multiple target languages, producing multilingual subtitles. This two-stage pipeline covers the complete requirement. Amazon Rekognition (A) analyzes video for objects, faces, and scenes -- it does not transcribe speech. Amazon Polly (C) converts text to speech, the opposite direction of what is needed. Amazon Textract (D) extracts text from images and documents, not from video audio tracks. See more: AWS Managed AI Services
Question 24 MEDIUM
A company is fine-tuning an Amazon Bedrock model on proprietary internal data. The security and compliance team requires that all fine-tuning data and resulting custom model weights are encrypted using the company's own encryption keys. Which AWS service enables this customer-controlled encryption?
AWS Key Management Service (KMS) enables customers to create and manage Customer Managed Keys (CMKs). Amazon Bedrock integrates with AWS KMS to encrypt fine-tuning datasets, custom model weights, and output artifacts using customer-controlled keys, satisfying data sovereignty and compliance requirements. AWS Shield (A) provides DDoS protection for AWS infrastructure. AWS Secrets Manager (C) stores and rotates application secrets such as API keys and database passwords. AWS Certificate Manager (D) provisions and manages TLS/SSL certificates for web applications. See more: AWS Security Services
Question 25 MEDIUM
Which of the following are characteristics of foundation models that distinguish them from traditional task-specific ML models? (Select TWO.)
Select all that apply
Foundation models have two key distinguishing characteristics: First, they are pre-trained on massive, diverse datasets using self-supervised learning objectives such as next-token prediction -- no manual labeling is required for pre-training. Second, they exhibit strong generalization, enabling adaptation to many different downstream tasks through zero-shot prompting, few-shot prompting, or fine-tuning with minimal task-specific data. Traditional ML models (B) are task-specific by design. The self-supervised pre-training objective (D) does not require labeled data -- labels are only needed during supervised fine-tuning for specific tasks. See more: AI/ML Fundamentals
Question 26 EASY
A company wants to convert call center audio recordings into searchable text transcripts. The recordings involve conversations between agents and customers, and the company needs to identify which speaker said what. Which Amazon Transcribe feature addresses this requirement?
Speaker diarization is an Amazon Transcribe feature that automatically detects multiple speakers in an audio recording and labels each segment with the speaker identifier who produced it. This creates structured transcripts such as 'Agent: Hello...' and 'Customer: Hi...' Custom vocabulary (A) adds domain-specific terms to improve transcription accuracy for specialized terminology. Vocabulary filtering (C) removes or masks specific words from transcripts for compliance. Automatic language identification (D) detects the spoken language without manual configuration but does not separate speakers. See more: AWS Managed AI Services
Question 27 MEDIUM
A company is using Amazon SageMaker to train a deep learning model. The training job is taking many hours due to the very large dataset size. The team wants to reduce training time without significantly impacting model accuracy. Which technique can reduce training time MOST effectively?
Distributed training splits the training workload across multiple compute instances using data parallelism, model parallelism, or a combination of both, significantly reducing wall-clock training time. Amazon SageMaker supports distributed training through the SageMaker Distributed Training Library. Increasing dropout rate (A) is a regularization technique that prevents overfitting but does not reduce training time. Decreasing model depth (C) reduces parameters and may speed up training but typically sacrifices model accuracy. L2 regularization (D) penalizes large model weights to prevent overfitting but does not accelerate training. See more: Amazon SageMaker
Question 28 EASY
A company is selecting a foundation model on Amazon Bedrock for a legal document review task. The documents can be up to 500 pages long and must be processed in a single API request to preserve full context. Which model characteristic is MOST critical for this use case?
Context window size determines the maximum number of tokens -- input and output combined -- that a model can process in a single request. Processing 500-page documents requires a very large context window (such as 200K tokens). Models with small context windows would require documents to be chunked into pieces, breaking cross-document reasoning. Number of parameters (A) relates to model capacity and capability but does not determine how much text can be processed at once. Training data recency (C) affects knowledge cutoff date. Inference latency (D) affects response speed but not document length capacity. See more: Bedrock & Generative AI
Question 29 MEDIUM
A data scientist is building a demand forecasting model using Amazon SageMaker. The model needs to forecast electricity consumption for the next 30 days using 3 years of historical hourly meter readings. Which SageMaker built-in algorithm is MOST suitable?
Amazon SageMaker DeepAR is a supervised learning algorithm specifically designed for probabilistic time-series forecasting using recurrent neural networks. It learns temporal patterns across multiple related time series simultaneously, producing accurate forecasts over specified future horizons. XGBoost (A) is a gradient boosting algorithm for tabular classification and regression -- it requires extensive feature engineering to capture time-series patterns. Linear Learner (B) is designed for linear classification and regression problems, not temporal forecasting. K-Means (D) is an unsupervised clustering algorithm with no forecasting capability. See more: Amazon SageMaker
Question 30 MEDIUM
A company wants to implement an AI governance framework. They need to document the intended use, limitations, training data provenance, and performance metrics for each ML model in a standardized format for regulatory compliance. Which AWS service supports this requirement?
Amazon SageMaker Model Cards provide a structured documentation template for ML models, including intended use cases, out-of-scope uses, model limitations, training data details, evaluation metrics, ethical considerations, and responsible AI practices. They serve as the standardized model governance artifact required for regulatory compliance. SageMaker Experiments (A) tracks training runs and hyperparameter configurations, not model governance documentation. AWS Audit Manager (C) collects evidence for compliance audits across AWS services but does not document ML model specifics. SageMaker Pipelines (D) automates ML workflows but does not produce governance documentation. See more: Amazon SageMaker
Question 31 MEDIUM
A company's AI hiring tool has been generating biased outputs that disadvantage certain demographic groups. The ML team wants to systematically detect and measure this bias across protected attributes. Which Amazon SageMaker feature is designed for this purpose?
Amazon SageMaker Clarify detects and measures statistical bias in both datasets and model predictions across different demographic groups. It computes bias metrics such as Difference in Positive Proportions (DPP) and Disparate Impact (DI) to identify unfair treatment across protected attributes. It also provides feature importance explanations using SHAP values to improve model transparency. SageMaker Model Monitor (A) monitors production models for data distribution changes and performance degradation, not specifically bias measurement. SageMaker Debugger (C) profiles training jobs for technical issues such as vanishing gradients. Feature Store (D) stores and serves ML features. See more: Amazon SageMaker
Question 32 MEDIUM
A company is building a conversational AI assistant using Amazon Bedrock. Users frequently ask questions about news events and developments that occurred after the foundation model's training cutoff date. Which technique BEST addresses this limitation?
RAG with a Knowledge Base that is continuously refreshed with current information enables the model to answer questions about recent events by retrieving up-to-date context at query time. This addresses the training cutoff limitation without expensive model retraining. Increasing temperature (A) affects output randomness and creativity, not knowledge recency. Weekly fine-tuning (C) is extremely expensive and slow -- the process cannot keep pace with the daily news cycle and still requires time to deploy. A larger model (D) has more parameters but its training cutoff remains the same; additional parameters do not provide new real-world knowledge. See more: Bedrock & Generative AI
Question 33 MEDIUM
A company wants to build a solution to automatically detect when customers in call center recordings express frustration or dissatisfaction during their conversations. Which combination of AWS services BEST achieves this?
Amazon Transcribe converts call center audio recordings into text transcripts. Amazon Comprehend then analyzes the transcripts for sentiment classification (positive, negative, neutral, mixed), detecting frustration and dissatisfaction signals at scale. Amazon Polly (B) converts text to speech, the opposite direction of what is needed. Amazon Rekognition (B) analyzes images and video for visual content, not audio sentiment. Amazon Textract (C) extracts text from scanned documents, not audio recordings. Amazon Kendra (C) is an enterprise search service for document retrieval. See more: AWS Managed AI Services
Question 34 MEDIUM
A company wants to use Amazon Bedrock to build a customer service chatbot. The chatbot must always respond in a formal professional tone and must never discuss competitor products under any circumstances. Which Amazon Bedrock features should the company configure to meet these requirements?
System prompts define the model's persona, communication style, and behavioral guidelines for every interaction, enforcing the formal tone requirement. Guardrails with denied topics configuration blocks the model from engaging with any content about competitors, enforcing that restriction reliably. Model fine-tuning and Provisioned Throughput (A) modify model weights and inference capacity but are not the primary mechanisms for enforcing behavioral tone or topic restrictions. Knowledge Bases and Agents (C) add retrieval and workflow capabilities. Continued pre-training and logging (D) add domain knowledge and audit trails, not behavioral controls. See more: Bedrock & Generative AI
Question 35 MEDIUM
A company trained an ML model to classify customer support tickets into three categories. The model achieves 95% accuracy on the training set but only 58% accuracy on new, unseen tickets. Which conclusion BEST describes this situation?
A large gap between high training accuracy (95%) and significantly lower test accuracy (58%) is the classic signature of overfitting. The model has memorized patterns specific to the training examples -- including noise -- rather than learning generalizable features that apply to new data. Underfitting (A) would result in poor performance on both training and test data, not just test data. Insufficient parameters (C) would cause underfitting, not a training-test accuracy gap. A higher learning rate (D) affects convergence speed and stability during training but does not resolve overfitting caused by memorization. See more: AI/ML Fundamentals
Question 36 HARD
A company wants to use Amazon Bedrock to enable employees to ask questions in plain English about data stored in a relational database, automatically generating and executing SQL queries without requiring employees to know SQL. Which Amazon Bedrock feature BEST supports this use case?
Amazon Bedrock Agents with action groups can connect to relational databases through AWS Lambda functions. The agent translates natural language questions into SQL queries, passes them to Lambda for execution against the database, and returns results as natural language responses. This enables true Text-to-SQL capability without requiring employees to know SQL. Knowledge Bases (A) are for document and vector search, not live relational database querying. Fine-tuning on SQL examples (C) improves SQL generation but does not give the model the ability to execute live queries against a database. Continued pre-training (D) adds domain knowledge to weights but cannot query live databases at inference time. See more: Bedrock & Generative AI
Question 37 MEDIUM
A company is building a foundation model (FM) for a specialized scientific domain. The company wants to teach the model broad knowledge about a new scientific field using a large collection of unlabeled research papers before task-specific adaptation. Which training approach is MOST appropriate?
Continued pre-training (also called domain-adaptive pre-training) extends the model's pre-training on large volumes of unlabeled domain-specific text, teaching the model the vocabulary, concepts, and patterns of a new scientific field without requiring labeled data. This is the appropriate technique when the goal is broad domain knowledge acquisition before any specific task. Fine-tuning with prompt-completion pairs (A) requires labeled data and targets specific task behavior rather than broad domain knowledge. RAG (C) retrieves documents at inference time but does not teach the model domain knowledge. Few-shot prompting (D) provides examples in the prompt but does not permanently update model knowledge. See more: Bedrock & Generative AI
Question 38 MEDIUM
A company must ensure that Amazon SageMaker training jobs process sensitive data in complete network isolation, with no ability to make outbound calls to the internet or other AWS services during training. Which solution meets this requirement?
SageMaker network isolation mode prevents training containers from making any outbound network calls during execution. When enabled, containers cannot reach the internet, other AWS services, or any external endpoints -- they can only read input data from and write output to Amazon S3 through VPC endpoints. SageMaker Experiments (A) is for tracking and comparing ML training runs, not for network isolation. Data encryption with KMS (C) protects data confidentiality but does not restrict network access during training. A restrictive IAM role (D) controls what AWS API calls the job can make but does not enforce network-level isolation from the internet. See more: AWS Security Services
Question 39 MEDIUM
A company is using an Amazon Bedrock base model for an internal application and has trained a custom fine-tuned version of the model to improve performance on their specific use case. What must the company do before they can use the custom fine-tuned model for inference through Amazon Bedrock?
Custom fine-tuned models in Amazon Bedrock cannot be invoked using on-demand pricing. To use a custom model for inference, the company must purchase Provisioned Throughput, which provides dedicated model units with either a 1-month or 6-month commitment term. Deploying to a SageMaker endpoint (B) is unnecessary because the model was fine-tuned within Amazon Bedrock and should be served from Bedrock. SageMaker Model Registry (C) is designed for models trained in SageMaker, not for Bedrock custom models. On-demand pricing (D) is not available for custom fine-tuned Bedrock models -- Provisioned Throughput is the required path. See more: Bedrock & Generative AI
Question 40 MEDIUM
An AI practitioner is building a model to generate images of professionals in various career fields. After reviewing the outputs, they discover the generated images consistently show gender and racial bias reflecting imbalances in the training dataset. Which technique BEST addresses this problem at its source?
Data augmentation for underrepresented groups addresses bias at its root by generating additional synthetic training examples for demographic groups that are underrepresented in the original dataset. Techniques such as flipping, rotation, color adjustment, and synthetic image generation can balance representation across professions, genders, and ethnicities before training. Watermark detection (A) identifies watermarks in images and is irrelevant to demographic bias. RAG (B) retrieves documents at inference time and is not applicable to image generation bias. Post-generation filtering (C) can screen outputs after the fact but does not fix the underlying model bias -- it treats the symptom rather than the cause. See more: AI Challenges & Responsibilities
Question 41 MEDIUM
A loan company is building a generative AI-based solution to offer new applicants discounts based on specific business criteria. The company wants to build and use an AI model responsibly to minimize bias that could negatively affect some customers. Which actions should the company take to meet these requirements? (Select TWO.)
Select all that apply
Two complementary approaches address bias: First, using diverse and representative training data ensures all demographic groups are proportionally included, preventing the model from favoring any particular group. Second, regular bias testing and fairness audits provide ongoing monitoring to detect and correct discriminatory patterns in model outputs over time. ROUGE (C) is a text summarization evaluation metric that measures overlap between generated and reference summaries -- it has nothing to do with bias detection or model accuracy in lending decisions. See more: AI Challenges & Responsibilities
Question 42 EASY
A company is training a foundation model (FM). The company wants to increase the accuracy of the model up to a specific acceptance level. Which solution will meet these requirements?
An epoch is one complete pass through the entire training dataset. Increasing epochs gives the model more opportunities to learn from the data, typically improving accuracy until the model converges. However, too many epochs can lead to overfitting, so training should stop when the target acceptance level is reached. Decreasing epochs (A) reduces training iterations, likely decreasing accuracy. Decreasing batch size (C) can affect training dynamics but doesn't directly or reliably increase accuracy the way more training iterations do. See more: AI/ML Fundamentals
Question 43 MEDIUM
A digital devices company wants to predict customer demand for memory hardware. The company does not have coding experience or knowledge of ML algorithms and needs to develop a data-driven predictive model. The company needs to perform analysis on internal data and external data. Which solution will meet these requirements?
Amazon SageMaker Canvas is the only truly no-code ML solution among the options. It provides a drag-and-drop interface where users upload datasets, select the prediction target, and Canvas automatically builds, trains, and evaluates multiple models. No coding, algorithm selection, or ML expertise is required. SageMaker built-in algorithms (A, B) require writing Python/R code and understanding ML concepts. Amazon Personalize Trending-Now (C) is a recommendation recipe designed for real-time trending product recommendations, not time-series demand forecasting. See more: Amazon SageMaker
Question 44 EASY
A large retailer receives thousands of customer support inquiries about products every day. The company wants to implement Agents for Amazon Bedrock. What are the key benefits of using Amazon Bedrock agents that could help this retailer?
Amazon Bedrock Agents enable end-to-end automation by decomposing customer requests into steps, invoking APIs (e.g., order lookup, return processing), querying knowledge bases for product information, and generating contextual responses -- all without manual intervention. This dramatically reduces the workload for customer support teams handling repetitive inquiries. The agent doesn't select between FMs (B) or call multiple FMs simultaneously (C) -- it uses one configured model and coordinates external actions through action groups and API integrations. See more: Bedrock & Generative AI
Question 45 MEDIUM
An AI practitioner is building a model to generate images of humans in various professions. The AI practitioner discovered that the input data is biased and that specific attributes affect the image generation and create bias in the model. Which technique will solve the problem?
Data augmentation for imbalanced classes addresses bias at its root by generating additional training examples for underrepresented groups. For example, if the dataset has few images of female engineers, augmentation techniques (rotation, flipping, color adjustment, synthetic generation) can create more examples, resulting in balanced representation across professions, genders, and ethnicities. Watermark detection (A) identifies watermarks in images, irrelevant to bias. RAG (B) retrieves information from knowledge bases, not applicable to image generation bias. Model monitoring (C) detects bias after deployment but doesn't fix the underlying training data imbalance. See more: AI Challenges & Responsibilities
Question 46 EASY
A retail company's website experiences unpredictable traffic spikes during flash sales. Which of the 6 advantages of cloud computing BEST addresses this challenge?
'Stop guessing capacity' directly addresses unpredictable traffic. With cloud computing, you can use Auto Scaling to automatically adjust resources based on actual demand, rather than over-provisioning for peak traffic or under-provisioning and crashing during spikes. Trading CAPEX for OPEX (A) relates to payment model. Economies of scale (C) relates to cost reduction from AWS's bulk purchasing. Going global (D) relates to geographic expansion. See more: AWS Cloud Computing
Question 47 EASY
According to the AWS Shared Responsibility Model, which of the following is the CUSTOMER's responsibility when using Amazon EC2?
EC2 is Infrastructure as a Service (IaaS). The customer is responsible for the guest operating system and everything above it, including OS patching, application security, and data protection. AWS is responsible for the underlying infrastructure: physical security (A), hypervisor (B), and hardware management (D). See more: AWS Cloud Computing
Question 48 EASY
A company wants to deploy a generative AI assistant that can answer employee questions using internal HR policies, project documents, and meeting notes stored in SharePoint and Amazon S3. Which AWS service is MOST appropriate?
Amazon Q Business is a fully managed enterprise AI assistant that connects to 40+ data sources including SharePoint and S3 via data connectors. It provides RAG-based retrieval from internal company data with permission-aware access control through IAM Identity Center. Bedrock (A) requires custom development. Lex (C) is for building conversational interfaces, not enterprise knowledge retrieval. Comprehend (D) does NLP analysis, not Q&A. See more: Amazon Q
Question 49 MEDIUM
An employee asks Amazon Q Business to 'Create a Jira ticket for the production outage on the payments service.' Which Q Business feature enables this action?
Plugins allow Amazon Q Business to actively interact with external systems like Jira, ServiceNow, and Salesforce -- creating, updating, or reading records. Data Connectors (A) are read-only integrations for RAG retrieval. Admin Controls (B) are guardrails for blocking topics. IAM Identity Center (D) handles user authentication, not external system actions. See more: Amazon Q
Question 50 MEDIUM
A SageMaker training job needs to read training data from an S3 bucket encrypted with AWS KMS. What IAM permissions must the SageMaker execution role have?
To access an encrypted S3 object, the SageMaker execution role needs both S3 permission to read the object AND KMS decrypt permission to decrypt the data. Without both permissions, the training job will fail. IAM roles are always required to authorize cross-service access in AWS. See more: AWS Security Services

Take Practice Test 2 →


Popular Posts