Generative AI LLM v1.0

Page:    1 / 4   
Exam contains 50 questions

In evaluating the transformer model for translation tasks, what is a common approach to assess its performance?

  • A. Analyzing the lexical diversity of the model’s translations compared to source texts.
  • B. Comparing the model’s output with human-generated translations on a standard dataset.
  • C. Evaluating the consistency of translation tone and style across different genres of text.
  • D. Measuring the syntactic complexity of the model’s translations against a corpus of professional translations.


Answer : B

Which metric is commonly used to evaluate machine-translation models?

  • A. Mean Absolute Error (MAE)
  • B. BLEU score
  • C. F1 score
  • D. Accuracy


Answer : B

You have developed a deep learning model for a recommendation system. You want to evaluate the performance of the model using A/B testing. What is the rationale for using A/B testing with deep learning model performance?

  • A. A/B testing ensures that the deep learning model is robust and can handle different variations of input data.
  • B. A/B testing allows for a controlled comparison between two versions of the model, helping to identify the version that performs better.
  • C. A/B testing methodologies integrate rationale and technical commentary from the designers of the deep learning model.
  • D. A/B testing helps in collecting comparative latency data to evaluate the performance of the deep learning model.


Answer : B

In the evaluation of Natural Language Processing (NLP) systems, what do ‘validity’ and ‘reliability’ imply regarding the selection of evaluation metrics?

  • A. Validity involves the metric’s ability to predict future trends in data, and reliability refers to its capacity to integrate with multiple data sources.
  • B. Validity ensures the metric accurately reflects the intended property to measure, while reliability ensures consistent results over repeated measurements.
  • C. Validity is concerned with the metric’s computational cost, while reliability is about its applicability across different NLP platforms.
  • D. Validity refers to the speed of metric computation, whereas reliability pertains to the metric’s performance in high-volume data processing.


Answer : B

What metrics would you use to evaluate the performance of a RAG workflow in terms of the accuracy of responses generated in relation to the input query? (Choose two.)

  • A. Generator latency
  • B. Retriever latency
  • C. Tokens generated per second
  • D. Response relevancy
  • E. Context precision


Answer : DE

What is the Open Neural Network Exchange (ONNX) format used for?

  • A. Representing deep learning models
  • B. Reducing training time of neural networks
  • C. Compressing deep learning models
  • D. Sharing neural network literature


Answer : A

You are in need of customizing your LLM via prompt engineering, prompt learning, or parameter efficient fine-tuning. Which framework helps you with all of these?

  • A. Nvidia TensorRT
  • B. Nvidia DALI
  • C. Nvidia Triton
  • D. Nvidia NeMO


Answer : D

Which of the following is a feature of the Nvidia Triton Inference Server?

  • A. Model quantization
  • B. Dynamic batching
  • C. Gradient clipping
  • D. Model pruning


Answer : B

Which of the following claims is correct about Tensor-RT and ONNX?

  • A. Tensor-RT is used for model deployment and ONNX is used for model interchange.
  • B. Tensor-RT is used for model deployment and ONNX is used for model creation.
  • C. Tensor-RT is used for model creation and ONNX is used for model interchange.
  • D. Tensor-RT is used for model creation and ONNX is used for model deployment.


Answer : A

Which of the following claims is correct about quantization in the context of Deep Learning? (Choose two.)

  • A. It only involves reducing the number of bits of the parameters.
  • B. Quantization might help in saving power and reducing heat production.
  • C. It leads to substantial loss of model accuracy.
  • D. It consists in removing a quantity of weights whose values are zero.
  • E. Helps reduce memory requirements and achieve better cache utilization.


Answer : BE

Which of the following optimizations are provided by TensorRT? (Choose two.)

  • A. Data augmentation
  • B. Variable learning rate
  • C. Multi-Stream Execution
  • D. Layer Fusion
  • E. Residual connections


Answer : CD

What is the purpose of the NVIDIA NGC catalog?

  • A. To provide a platform for testing and debugging software applications.
  • B. To provide a platform for developers to collaborate and share software development projects.
  • C. To provide a marketplace for buying and selling software development tools and resources.
  • D. To provide a curated collection of GPU-optimized AI and data science software.


Answer : D

Imagine you are training an LLM consisting of billions of parameters and your training dataset is significantly larger than the available RAM in your system. Which of the following would be an alternative?

  • A. Using the GPU memory to extend the RAM capacity for storing the dataset and move the dataset in and out of the GPU, using the PCI bandwidth possibly.
  • B. Using a memory-mapped file that allows the library to access and operate on elements of the dataset without needing to fully load it into memory.
  • C. Discarding the excess of data and pruning the dataset to the capacity of the RAM, resulting in reduced latency during inference.
  • D. Eliminating sentences that are syntactically different by semantically equivalent, possibly reducing the risk of the model hallucinating as it is trained to get to the point.


Answer : B

When implementing data parallel training, which of the following considerations needs to be taken into account?

  • A. The model weights are synced across all processes/devices only at the end of every epoch.
  • B. A master-worker method for syncing the weights across different processes is desirable due to its scalability.
  • C. A ring all-reduce is an efficient algorithm for syncing the weights across different processes/devices.
  • D. The model weights are kept independent for as long as possible, increasing the model exploration.


Answer : C

“Hallucinations” is a term coined to describe when LLM models produce what?

  • A. Outputs are only similar to the input data.
  • B. Images from a prompt description.
  • C. Correct sounding results that are wrong.
  • D. Grammatically incorrect or broken outputs.


Answer : C

Page:    1 / 4   
Exam contains 50 questions

Talk to us!


Have any questions or issues ? Please dont hesitate to contact us

Certlibrary.com is owned by MBS Tech Limited: Room 1905 Nam Wo Hong Building, 148 Wing Lok Street, Sheung Wan, Hong Kong. Company registration number: 2310926
Certlibrary doesn't offer Real Microsoft Exam Questions. Certlibrary Materials do not contain actual questions and answers from Cisco's Certification Exams.
CFA Institute does not endorse, promote or warrant the accuracy or quality of Certlibrary. CFA® and Chartered Financial Analyst® are registered trademarks owned by CFA Institute.
Terms & Conditions | Privacy Policy | Amazon Exams | Cisco Exams | CompTIA Exams | Databricks Exams | Fortinet Exams | Google Exams | Microsoft Exams | VMware Exams