top of page
  • Writer's pictureKaty Morgan

Introduction to Microsoft Certified - Azure Data Scientist Associate Exam

The Microsoft DP-100 Exam is challenging and thorough preparation is essential for success. This exam study guide is designed to help you prepare for the Designing and Implementing a Data Science Solution on Azure certification exam. It contains a detailed list of the topics covered on the Professional exam, as well as a detailed list of preparation resources. These study guides for the Designing and Implementing a Data Science Solution on Microsoft Azure will help guide you through the study process for your certification.

DP-100 pdf, DP-100 questions, DP-100 exam guide, DP-100 practice test, DP-100 books, DP-100 Syllabus, DP-100

DP-100 Designing and Implementing a Data Science Solution on Microsoft Azure Exam Summary


Exam Name: Designing and Implementing a Data Science Solution on Microsoft Azure

Exam Code: DP-100

Exam Price: $550 (USD)

Duration: 120 mins

Number of Questions: 40-60

Passing Score: 700 / 1000

Schedule Exam: Pearson VUE


Exam Syllabus: DP-100 Microsoft Certified - Azure Data Scientist Associate


1. Manage Azure resources for machine learning (25-30%)


Create an Azure Machine Learning workspace

- create an Azure Machine Learning workspace

- configure workspace settings

- manage a workspace by using Azure Machine Learning studio

Manage data in an Azure Machine Learning workspace

- select Azure storage resources

- register and maintain datastores

- create and manage datasets

Manage compute for experiments in Azure Machine Learning

- determine the appropriate compute specifications for a training workload

- create compute targets for experiments and training

- configure Attached Compute resources including Azure Databricks

- monitor compute utilization

Implement security and access control in Azure Machine Learning

- determine access requirements and map requirements to built-in roles

- create custom roles

- manage role membership

- manage credentials by using Azure Key Vault

Set up an Azure Machine Learning development environment

- create compute instances

- share compute instances

- access Azure Machine Learning workspaces from other development environments

Set up an Azure Databricks workspace

- create an Azure Databricks workspace

- create an Azure Databricks cluster

- create and run notebooks in Azure Databricks

- link and Azure Databricks workspace to an Azure Machine Learning workspace


2. Run Experiments and Train Models (20-25%)


Create models by using the Azure Machine Learning Designer

- create a training pipeline by using Azure Machine Learning designer

- ingest data in a designer pipeline

- use designer modules to define a pipeline data flow

- use custom code modules in designer

Run model training scripts

- create and run an experiment by using the Azure Machine Learning SDK

- configure run settings for a script

- consume data from a dataset in an experiment by using the Azure Machine Learning SDK

- run a training script on Azure Databricks compute

- run code to train a model in an Azure Databricks notebook

Generate metrics from an experiment run

- log metrics from an experiment run

- retrieve and view experiment outputs

- use logs to troubleshoot experiment run errors

- use MLflow to track experiments

- track experiments running in Azure Databricks

Use Automated Machine Learning to create optimal models

- use the Automated ML interface in Azure Machine Learning studio

- use Automated ML from the Azure Machine Learning SDK

- select pre-processing options

- select the algorithms to be searched

- define a primary metric

- get data for an Automated ML run

- retrieve the best model

Tune hyperparameters with Azure Machine Learning

- select a sampling method

- define the search space

- define the primary metric

- define early termination options

- find the model that has optimal hyperparameter values


3. Deploy and operationalize machine learning solutions (35-40%)


Select compute for model deployment

- consider security for deployed services

- evaluate compute options for deployment

Deploy a model as a service

- configure deployment settings

- deploy a registered model

- deploy a model trained in Azure Databricks to an Azure Machine Learning endpoint

- consume a deployed service

- troubleshoot deployment container issues

Manage models in Azure Machine Learning

- register a trained model

- monitor model usage

- monitor data drift

Create an Azure Machine Learning pipeline for batch inferencing

- configure a ParallelRunStep

- configure compute for a batch inferencing pipeline

- publish a batch inferencing pipeline

- run a batch inferencing pipeline and obtain outputs

- obtain outputs from a ParallelRunStep

Publish an Azure Machine Learning designer pipeline as a web service

- create a target compute resource

- configure an Inference pipeline

- consume a deployed endpoint

Implement pipelines by using the Azure Machine Learning SDK

- create a pipeline

- pass data between steps in a pipeline

- run a pipeline

- monitor pipeline runs

Apply ML Ops practices

- trigger an Azure Machine Learning pipeline from Azure DevOps

- automate model retraining based on new data additions or data changes

- refactor notebooks into scripts

- implement source control for scripts


4. Implement Responsible ML (5-10%)


Use model explainers to interpret models

- select a model interpreter

- generate feature importance data

Describe fairness considerations for models

- evaluate model fairness based on prediction disparity

- mitigate model unfairness

Describe privacy considerations for data

- describe principles of differential privacy

- specify acceptable levels of noise in data and the effects on privacy


Microsoft DP-100 Certification Sample Questions and Answers


To make you familiar with Designing and Implementing a Data Science Solution on Microsoft Azure (DP-100) certification exam structure, we have prepared this sample question set. We suggest you try our Sample Questions for Designing and Implementing a Data Science Solution on Azure DP-100 Certification to test your understanding of Microsoft DP-100process with the real Microsoft certification exam environment.


DP-100 Designing and Implementing a Data Science Solution on Microsoft Azure Sample Questions:-


01. You use the Two-Class Neural Network module in Azure Machine Learning Studio to build a binary classification model. You use the Tune Model Hyperparameters module to tune accuracy for the model.

You need to configure the Tune Model Hyperparameters module. Which two values should you use?

Each correct answer presents part of the solution. NOTE: Each correct selection is worth one point.

a) Number of hidden nodes

b) Learning Rate

c) The type of the normalizer

d) Number of learning iterations

e) Hidden layer specification


02. You are analyzing a dataset by using Azure Machine Learning Studio. You need to generate a statistical summary that contains the p-value and the unique count for each feature column.

Which two modules can you use?

Each correct answer presents a complete solution. NOTE: Each correct selection is worth one point.

a) Computer Linear Correlation

b) Export Count Table

c) Execute Python Script

d) Convert to Indicator Values

e) Summarize Data


03. You train a machine learning model. You must deploy the model as a real-time inference service for testing.

The service requires low CPU utilization and less than 48 MB of RAM. The compute target for the deployed service must initialize automatically while minimizing cost and administrative overhead.

Which compute target should you use?

a) Azure Container Instance (ACI)

b) attached Azure Databricks cluster

c) Azure Kubernetes Service (AKS) inference cluster

d) Azure Machine Learning compute cluster


04. You manually define security rules to isolate your Azure Machine Learning training processes within an Azure Virtual Network. You need to ensure that only required communications from the Internet are allowed.

Which three actions should you perform?

Each correct answer presents part of the solution. Choose the correct answers.

a) Define an inbound security rule.

b) Ensure the source service tag is set to AzureMachineLearning.

c) Ensure the source service tag is set to BatchNodeManagement.

d) Configure a source port range of 29876-29877.

e) Create an outbound security rule.

f) Configure a destination port range of 22.


05. You plan to use a Deep Learning Virtual Machine (DLVM) to train deep learning models using Compute Unified Device Architecture (CUDA) computations.

You need to configure the DLVM to support CUDA. What should you implement?

a) Solid State Drives (SSD)

b) Computer Processing Unit (CPU) speed increase by using overclocking

c) Graphic Processing Unit (GPU)

d) High Random Access Memory (RAM) configuration

e) Intel Software Extensions (Intel SGX) technology


06. You are building a regression model for estimating the number of calls during an event. You need to determine whether the feature values achieve the conditions to build a Poisson regression model.

Which two conditions must the feature set contain?

Each correct answer presents part of the solution. NOTE: Each correct selection is worth one point.

a) The label data must be a negative value.

b) The label data must be whole numbers.

c) The label data must be non-discrete.

d) The label data must be a positive value.

e) The label data can be positive or negative.


07. You use Azure Machine Learning designer to publish an inference pipeline as a web service. During deployment, you create a service principal (SP) and configure authentication.

You need to use the SP while consuming the endpoint. What should you do?

Choose the correct answer.

a) Use the client secret to retrieve an authentication token.

b) Use the regen_key method to regenerate the primary key.

c) Use the AciWebservice.deploy_configuration to set auth_enabled to True.

d) Use the get_keys method to retrieve authentication keys.


08. You retrain an existing model. You need to register the new version of a model while keeping the current version of the model in the registry. What should you do?

a) Register a model with a different name from the existing model and a custom property named version with the value 2.

b) Register the model with the same name as the existing model.

c) Save the new model in the default datastore with the same name as the existing model. Do not register the new model.

d) Delete the existing model and register the new one with the same name.


09. You plan to provision an Azure Machine Learning Basic edition workspace for a data science project. You need to identify the tasks you will be able to perform in the workspace.

Which three tasks will you be able to perform?

Each correct answer presents a complete solution. NOTE: Each correct selection is worth one point.

a) Create a Compute Instance and use it to run code in Jupyter notebooks.

b) Create an Azure Kubernetes Service (AKS) inference cluster.

c) Use the designer to train a model by dragging and dropping pre-defined modules.

d) Create a tabular dataset that supports versioning.

e) Use the Automated Machine Learning user interface to train a model.


10. You have access to structured data used in a previous machine learning project. The data is not registered in a dataset.

You need to use Azure Machine Learning SDK to use the data directly in a machine learning script. What should you do?

Choose the correct answer

a) Use Python to create a TabularDataset. Specify the path to the data.

b) Use the get_context method from the Run class to load the run from the remote environment.

c) Use the register method from the Dataset class to create a new dataset version.

d) Use the from_files method to create a dataset and register the dataset in your workspace.


Answers:-


Answer 1:- b, d

Answer 2:- b, e

Answer 3:- a

Answer 4:- a, c, d

Answer 5:- c

Answer 6:- b, d

Answer 7:- a

Answer 8:- a

Answer 9:- a, b, c, d

Answer 10:- a

bottom of page