Summer Sale Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: ecus65

Microsoft DP-100 - Designing and Implementing a Data Science Solution on Azure

Page: 4 / 5
Total 476 questions

You have a dataset that contains over 150 features. You use the dataset to train a Support Vector Machine (SVM) binary classifier.

You need to use the Permutation Feature Importance module in Azure Machine Learning Studio to compute a set of feature importance scores for the dataset.

In which order should you perform the actions? To answer, move all actions from the list of actions to the answer area and arrange them in the correct order.

You use Azure Machine Learning to implement hyperparameter tuning with a Bandit early termination policy.

The policy uses a slack_factor set to 01. an evaluation interval set to 1, and an evaluation delay set to b.

You need to evaluate the outcome of the early termination policy

What should you evaluate? To answer, select the appropriate options m the answer area.

NOTE: Each correct selection is worth one point.

You are using the Azure Machine Learning designer to transform a dataset by using an Execute Python Script component and custom code.

You need to define the method signature for the Execute Python Script component and return value type.

What should you define? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

You use the Azure Machine Learning designer to create and run a training pipeline. You then create a real-time inference pipeline.

You must deploy the real-time inference pipeline as a web service.

What must you do before you deploy the real-time inference pipeline?

A.

Run the real-time inference pipeline.

B.

Create a batch inference pipeline.

C.

Clone the training pipeline.

D.

Create an Azure Machine Learning compute cluster.

You create a script that trains a convolutional neural network model over multiple epochs and logs the validation loss after each epoch. The script includes arguments for batch size and learning rate.

You identify a set of batch size and learning rate values that you want to try.

You need to use Azure Machine Learning to find the combination of batch size and learning rate that results in the model with the lowest validation loss.

What should you do?

A.

Run the script in an experiment based on an AutoMLConfig object

B.

Create a PythonScriptStep object for the script and run it in a pipeline

C.

Use the Automated Machine Learning interface in Azure Machine Learning studio

D.

Run the script in an experiment based on a ScriptRunConfig object

E.

Run the script in an experiment based on a HyperDriveConfig object

You are creating a new experiment in Azure Machine Learning Studio. You have a small dataset that has missing values in many columns. The data does not require the application of predictors for each column. You plan to use the Clean Missing Data module to handle the missing data.

You need to select a data cleaning method.

Which method should you use?

A.

Synthetic Minority Oversampling Technique (SMOTE)

B.

Replace using MICE

C.

Replace using; Probabilistic PCA

D.

Normalization

You have an Azure Machine Learning workspace that includes an AmICompute cluster and a batch endpoint. You clone a repository that contains an MLflow model to your local computer. You need to ensure that you can deploy the model to the batch endpoint.

Solution: Create a data asset in the workspace.

Does the solution meet the goal?

A.

Yes

B.

No

You use Azure Machine Learning designer to create a training pipeline for a regression model.

You need to prepare the pipeline for deployment as an endpoint that generates predictions asynchronously for a dataset of input data values.

What should you do?

A.

Clone the training pipeline.

B.

Create a batch inference pipeline from the training pipeline.

C.

Create a real-time inference pipeline from the training pipeline.

D.

Replace the dataset in the training pipeline with an Enter Data Manually module.

You are moving a large dataset from Azure Machine Learning Studio to a Weka environment.

You need to format the data for the Weka environment.

Which module should you use?

A.

Convert to CSV

B.

Convert to Dataset

C.

Convert to ARFF

D.

Convert to SVMLight

You manage an Azure Machine Learning won pace named workspace 1 by using the Python SDK v2. You create a Gene-al Purpose v2 Azure storage account named mlstorage1. The storage account includes a pulley accessible container name micOTtalnerl. The container stores 10 blobs with files in the CSV format.

You must develop Python SDK v2 code to create a data asset referencing all blobs in the container named mtcontamer1.

You need to complete the Python SDK v2 code.

How should you complete the code? To answer select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.