Become Microsoft Certified with updated DP-100 exam questions and correct answers
You have been tasked with creating a new Azure pipeline via the Machine Learning designer. You have to makes sure that the pipeline trains a model using data in a comma-separated values (CSV) file that is published on a website. A dataset for the file for this file does not exist. Data from the CSV file must be ingested into the designer pipeline with the least amount of administrative effort as possible. Which of the following actions should you take?
You plan to run a Python script as an Azure Machine Learning experiment. The script must read files from a hierarchy of folders. The files will be passed to the script as a dataset argument. You must specify an appropriate mode for the dataset argument. Which two modes can you use? Each correct answer presents a complete solution. NOTE: Each correct selection is worth one point.
You are implementing a machine learning model to predict stock prices. The model uses a PostgreSQL database and requires GPU processing. You need to create a virtual machine that is pre-configured with the required tools. What should you do?
Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution. After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen. You are analyzing a numerical dataset which contains missing values in several columns. You must clean the missing values using an appropriate operation without affecting the dimensionality of the feature set. You need to analyze a full dataset to include all values. Solution: Replace each missing value using the Multiple Imputation by Chained Equations (MICE) method. Does the solution meet the goal?
You create an Azure Machine Learning workspace. The workspace contains a dataset named
sample.dataset, a compute instance, and a compute cluster. You must create a two-stage pipeline
that will prepare data in the dataset and then train and register a model based on the prepared data.
The first stage of the pipeline contains the following code:
You need to identify the location containing the output of the first stage of the script that you can use
as input for the second stage. Which storage location should you use?
© Copyrights DumpsCertify 2026. All Rights Reserved
We use cookies to ensure your best experience. So we hope you are happy to receive all cookies on the DumpsCertify.