NCA-GENM試験無料問題集「NVIDIA Generative AI Multimodal 認定」
You are tasked with optimizing a large multimodal AI model for deployment on edge devices with limited computational resources. Which combination of techniques would provide the BEST trade-off between model accuracy and inference speed? (Select TWO)
正解:B,E
解答を投票する
解説: (GoShiken メンバーにのみ表示されます)
Consider the following Python code snippet using PyTorch, designed to combine text and image embeddings before feeding them into a transformer. Assume 'text_embedding' has shape '(batch_size, seq_len, hidden_dim)' and 'image_embedding' has shape '(batch_size, image_features)'. Which of the following code snippets MOST correctly combines these embeddings for a multimodal transformer input?
正解:B
解答を投票する
解説: (GoShiken メンバーにのみ表示されます)
You have a multimodal model that processes images and text, and you want to deploy it on an edge device with limited computational resources. Which of the following hardware acceleration strategies would be MOST effective in improving the model's inference speed on the edge device?
正解:A,C
解答を投票する
解説: (GoShiken メンバーにのみ表示されます)
You have a multimodal model combining video and text data for action recognition. The model performs well on standard datasets but struggles with videos containing unusual camera angles or lighting conditions. Which data augmentation strategy would be MOST effective in improving the model's robustness?
正解:B
解答を投票する
解説: (GoShiken メンバーにのみ表示されます)
Consider the following code snippet using a hypothetical Generative A1 library. This code is intended to generate an image from a text prompt and then refine it based on a user-provided style image. However, it's not producing the desired results. What is the MOST likely cause of the issue?
正解:C
解答を投票する
解説: (GoShiken メンバーにのみ表示されます)
You are working on a project to classify images of different types of flowers. You have a relatively small dataset (around 500 images per class). Which of the following techniques would be the MOST effective to improve the performance of your image classifier, considering the limited data?
正解:E
解答を投票する
解説: (GoShiken メンバーにのみ表示されます)
You are tasked with optimizing a multimodal A1 model that processes both image and text data for generating image captions. The model exhibits slow inference times, particularly when handling high-resolution images. Which of the following optimization strategies would be MOST effective in reducing inference latency, considering the NVIDIA ecosystem?
正解:D
解答を投票する
解説: (GoShiken メンバーにのみ表示されます)
You are building a real-time image captioning system using a Transformer model. You observe significant latency issues when generating captions for high-resolution images. Which optimization strategies would be most effective in reducing the latency without significantly sacrificing caption quality? (Select all that apply)
正解:B,C,D
解答を投票する
解説: (GoShiken メンバーにのみ表示されます)
You are using NeMo to fine-tune a pre-trained language model for a specific text generation task. You want to implement a custom data augmentation technique to improve the model's robustness. Which of the following approaches is most appropriate for integrating your custom augmentation within the NeMo framework?
正解:D
解答を投票する
解説: (GoShiken メンバーにのみ表示されます)
You are developing an Avatar Cloud Engine (ACE) application for a virtual assistant that needs to generate realistic facial expressions based on user emotions detected from text. Which ACE microservice would be most directly responsible for this functionality?
正解:D
解答を投票する
解説: (GoShiken メンバーにのみ表示されます)
You are building a multimodal generative A1 model that combines text, images, and audio. You notice that the model performs well on text and images but struggles with audio, particularly in noisy environments. Which of the following strategies would be MOST effective in improving the model's performance with audio data?
正解:A,C
解答を投票する
解説: (GoShiken メンバーにのみ表示されます)