Agentic AI v1.0

Page:    1 / 9   
Exam contains 121 questions

You’re employing an LLM to automate the generation of email responses for a customer service team. The generated responses frequently miss the mark, failing to address the customer’s underlying concerns.
What’s the most crucial element to add to the prompt to enhance the quality of the email responses?

  • A. Instructing the LLM with a detailed prompt containing instructions on how to format and compose the response in an easy-to-understand structure.
  • B. Instructing the LLM to use a simple template for all email replies before generating a response.
  • C. Instructing the LLM to “understand the customer’s issue” before generating a response.
  • D. Instructing the LLM to provide a response that “is the most helpful” before generating a response.


Answer : A

After a series of adjustments in a supply chain agentic system, the agent has dramatically reduced shipping times and minimized costs, but the team is receiving a high volume of complaints from customers regarding delayed deliveries.
Which metric is MOST important to prioritize when investigating this situation?

  • A. The agent’s ability to predict future demand fluctuations, as accurate forecasting is crucial for effective logistics.
  • B. The total cost savings achieved through the agent’s optimization, which represents a significant financial benefit.
  • C. The percentage of delivery times that fall within the acceptable delay window, considering customer satisfaction as a key factor.
  • D. The agent’s adherence to the prescribed delivery schedules, as it’s demonstrably improving efficiency.


Answer : C

A recently deployed Agentic AI system designed for automated incident response within a cloud infrastructure has been consistently failing to identify and resolve ‘high-priority’ alerts – specifically, those related to increased CPU utilization across several virtual machines. Initial logs show the agent is primarily focusing on alerts with related network traffic spikes, ignoring the CPU metrics.
What is the most appropriate initial step for a senior Agentic AI engineer to take to resolve this issue, considering the system’s reliance on benchmarking and iterative improvement?

  • A. Review the agent’s evaluation framework, focusing on the defined benchmarks used to assess its response efficiency and impact on overall system performance.
  • B. Replace the agent’s underlying AI model with a more powerful, general-purpose machine learning engine as a first step in investigating current benchmarks.
  • C. Implement a new synthetic data set containing a wide variety of CPU load profiles to train the agent’s decision-making model.
  • D. Review the agent’s sensitivity thresholds, focusing on CPU utilization alerts to maximize detection accuracy.


Answer : A

A team is evaluating multiple versions of an AI agent designed for customer support. They want to identify which version completes tasks more efficiently, responds accurately, and improves over time using user feedback.
Which practice is most important to ensure continuous refinement and optimal performance of the AI agent?

  • A. Comparing agents on isolated tasks without standardized benchmarking pipelines
  • B. Relying solely on offline benchmarks without incorporating live user feedback during tuning
  • C. Implementing an evaluation framework that quantifies task efficiency and incorporates human-in-the-loop feedback
  • D. Tuning model parameters once before deployment to maximize initial accuracy


Answer : C

When analyzing inconsistent performance across a fleet of customer service agents handling similar queries, which evaluation approach most effectively identifies root causes and optimization opportunities?

  • A. Assess performance data from recently improved agents and highlight strong results, using outcome comparisons to identify areas with the greatest impact on service quality.
  • B. Average performance metrics across all agents as this will smooth individual variations, query distribution differences, and temporal factors affecting agent behavior and accuracy.
  • C. Deploy stratified evaluation sampling across agent variants, query complexity levels, and temporal patterns while tracking decision paths using comparative analytics.
  • D. Review performance across both high- and low-accuracy agent groups, comparing case outcomes and identifying patterns contributing to top and bottom results.


Answer : C

You are using an LLM-as-a-Judge to evaluate a RAG pipeline.
What is the primary benefit of synthetically generating question-answer pairs, rather than relying solely on human-created test cases?

  • A. Synthetically generated questions are more challenging and reveal deeper flaws in the RAG pipeline.
  • B. Synthetic generation eliminates the need for any human validation of the RAG pipeline’s output.
  • C. Synthetically generated answers are inherently more accurate than those produced by the LLM.
  • D. Synthetic generation allows for systematic testing of the RAG pipeline across a wider range of scenarios and query types.


Answer : D

Your agent is generating inconsistent and contradictory statements.
Which approach would be most suitable to improve the agent’s output?

  • A. Employing Reflexion
  • B. Increasing the number of generated plans
  • C. Using Decomposition-First Planning
  • D. Decreasing the length of prompts


Answer : A

You’re utilizing an LLM to translate complex technical documentation into multiple languages. The translations often lack nuance and fail to capture the original intent.
What’s the most effective strategy for improving the quality of the translations?

  • A. Providing the LLM with a glossary of key terms, concepts in all languages and the dataset of previously translated text.
  • B. Training the LLM on a dataset of translated texts.
  • C. Providing the LLM with guidance to “translate the documents” without additional guidance, so it can use trained knowledge.
  • D. Providing the LLM with guidance to translate “with high accuracy” without additional guidance, so it can use trained knowledge.


Answer : A

An e-commerce platform is implementing an AI-powered customer support system that handles inquiries ranging from simple FAQ responses to complex product recommendations and technical troubleshooting. The system experiences unpredictable traffic patterns with sudden spikes during sales events and varying complexity requirements. Simple questions comprise the majority of requests but require minimal compute, while complex product recommendations need sophisticated reasoning. The company wants to optimize costs while maintaining service quality across all query types.
Which approach would provide the MOST cost-optimized scaling strategy for this variable-workload, mixed-complexity environment?

  • A. Deploy specialized NVIDIA NIM microservices using a single large model configuration that handles all agent functions on high-capacity GPUs, with auto-scaling infrastructure that maintains constant resource allocation across all traffic patterns.
  • B. Deploy specialized NVIDIA NIM microservices on CPU-optimized infrastructure with auto-scaling capabilities to minimize hardware costs, while accepting longer inference times for cost optimization benefits.
  • C. Deploy specialized NVIDIA NIM microservices with an LLM router to dynamically route requests to appropriate models based on complexity, combined with auto-scaling infrastructure that scales different model types independently.
  • D. Deploy multiple specialized NVIDIA NIM microservices with identical high-capacity models across all available GPUs, implementing auto-scaling infrastructure without request complexity differentiation or dynamic model selection capabilities.


Answer : C

A technology startup is preparing to launch an AI agent platform to serve clients with unpredictable usage patterns. They face periods of high user activity and low demand, so their deployment approach must minimize wasted resources during slow times and automatically allocate more resources during busy periods – all while keeping operational costs reasonable.
Given these requirements, which deployment strategy most effectively ensures both cost-effectiveness and adaptability for scaling agentic AI systems?

  • A. Scheduling periodic manual reviews to increase or decrease infrastructure based on predicted user numbers
  • B. Monitoring system logs for usage patterns and making infrastructure changes after monthly analysis
  • C. Using fixed-size virtual machine clusters to guarantee consistent resource allocation at all times
  • D. Implementing autoscaling policies in a container orchestration environment to automatically adjust resources according to workload changes


Answer : D

When evaluating a multi-agent customer service system experiencing unpredictable scaling costs and performance bottlenecks during peak hours, which analysis approaches effectively identify optimization opportunities for both infrastructure efficiency and service reliability? (Choose two.)

  • A. Maintain consistent resource allocation across all service hours, for a more precise view of baseline traffic impact on long-term infrastructure efficiency.
  • B. Scale agent infrastructure based on aggregate performance trends, using system-wide monitoring tools to identify broader optimization patterns across resources.
  • C. Deploy agents with configurable scaling workflows, allowing analysis of resource adjustment strategies and their effects on service stability during variable demand periods.
  • D. Deploy distributed tracing with cost attribution per agent type, correlating resource consumption with business value metrics to identify optimization opportunities in agent deployment strategies.
  • E. Implement comprehensive workload profiling using NVIDIA Nsight to analyze GPU utilization patterns, identify underutilized resources, and optimize batch sizing for dynamic scaling with Kubernetes HPA.


Answer : DE

When analyzing throughput bottlenecks in a multi-modal agent processing text, images, and audio, which Triton configuration evaluations identify optimization opportunities? (Choose two.)

  • A. Analyze model ensemble pipelines for sequential dependencies, identify parallelization opportunities, and optimize inter-model data transfer using Triton’s scheduler.
  • B. Profile GPU memory allocation patterns across modalities, implement model instance batching strategies, and tune concurrency limits to maximize utilization.
  • C. Deploy each modality on separate Triton instances, allowing Triton to automatically manage ensemble coordination, shared memory usage, and pipeline integration.
  • D. Use a single model instance per GPU, allowing Triton to automatically optimize concurrency, batching, and multi-instance settings for throughput scaling.


Answer : AB

When analyzing performance bottlenecks in a multi-modal agent processing customer support tickets with text, images, and voice inputs, which evaluation approach most effectively identifies optimization opportunities?

  • A. Measure total response time as this analyzes aggregated performance trends across modalities, model loading times, and opportunities for parallel execution.
  • B. Profile end-to-end latency across modalities, measure model switching overhead, analyze batch processing opportunities, and evaluate Triton’s dynamic batching for multi-modal workloads.
  • C. Optimize each modality independently using dedicated profiling of cross-modal interactions, shared resource constraints, and pipeline execution strategies.
  • D. Extend evaluation to accuracy and quality metrics, incorporating resource usage patterns, latency observations, and their impact on user experience.


Answer : B

What benefits does a Kubernetes deployment offer over Slurm?

  • A. Kubernetes provides autoscaling, auto-restarts, dynamic task scheduling, error isolation with containers, and integrated monitoring.
  • B. Kubernetes is the best option for both training and inference, offering advantages for resource management and workload visibility over traditional HPC schedulers like Slurm.
  • C. Kubernetes is more optimized for batch jobs to achieve high throughput, and also provides for monitoring and failover in large-scale workloads.


Answer : A

A company plans to launch a multi-agent system that must serve thousands of users simultaneously. The team needs to ensure the system remains reliable, scales efficiently as demand increases, and operates in a cost-effective manner.
Which approach is most effective for achieving robust and scalable deployment of an agentic AI system in production?

  • A. Running agents without load balancing to reduce infrastructure complexity and achieve robust and scalable deployment of an agentic system
  • B. Establishing a continuous monitoring framework to track system performance and adapt resources as usage patterns evolve
  • C. Deploying all agents on a single server with ongoing performance monitoring to maximize hardware utilization
  • D. Orchestrating agents using containerization platforms, combined with load balancing and ongoing performance monitoring


Answer : D

Page:    1 / 9   
Exam contains 121 questions

Talk to us!


Have any questions or issues ? Please dont hesitate to contact us

Certlibrary.com is owned by MBS Tech Limited: Room 1905 Nam Wo Hong Building, 148 Wing Lok Street, Sheung Wan, Hong Kong. Company registration number: 2310926
Certlibrary doesn't offer Real Microsoft Exam Questions. Certlibrary Materials do not contain actual questions and answers from Cisco's Certification Exams.
CFA Institute does not endorse, promote or warrant the accuracy or quality of Certlibrary. CFA® and Chartered Financial Analyst® are registered trademarks owned by CFA Institute.
Terms & Conditions | Privacy Policy | Amazon Exams | Cisco Exams | CompTIA Exams | Databricks Exams | Fortinet Exams | Google Exams | Microsoft Exams | VMware Exams