The Microsoft DP-100 Exam is challenging and thorough preparation is essential for success. This exam study guide is designed to help you prepare for the Designing and Implementing a Data Science Solution on Azure certification exam. It contains a detailed list of the topics covered on the Professional exam, as well as a detailed list of preparation resources. These study guides for the Designing and Implementing a Data Science Solution on Microsoft Azure will help guide you through the study process for your certification.
DP-100 Designing and Implementing a Data Science Solution on Microsoft Azure Exam Summary
● Exam Name: Designing and Implementing a Data Science Solution on Microsoft Azure
● Exam Code: DP-100
● Exam Price: $550 (USD)
● Duration: 120 mins
● Number of Questions: 40-60
● Passing Score: 700 / 1000
● Books / Training: DP-100T01-A: Designing and Implementing a Data Science Solution on Azure
● Schedule Exam: Pearson VUE
● Sample Questions: Microsoft Designing and Implementing a Data Science Solution on Azure Sample Questions
● Recommended Practice: Microsoft DP-100 Certification Practice Exam
Exam Syllabus: DP-100 Microsoft Certified - Azure Data Scientist Associate
1. Manage Azure resources for machine learning (25-30%)
Create an Azure Machine Learning workspace
- create an Azure Machine Learning workspace
- configure workspace settings
- manage a workspace by using Azure Machine Learning studio
Manage data in an Azure Machine Learning workspace
- select Azure storage resources
- register and maintain datastores
- create and manage datasets
Manage compute for experiments in Azure Machine Learning
- determine the appropriate compute specifications for a training workload
- create compute targets for experiments and training
- configure Attached Compute resources including Azure Databricks
- monitor compute utilization
Implement security and access control in Azure Machine Learning
- determine access requirements and map requirements to built-in roles
- create custom roles
- manage role membership
- manage credentials by using Azure Key Vault
Set up an Azure Machine Learning development environment
- create compute instances
- share compute instances
- access Azure Machine Learning workspaces from other development environments
Set up an Azure Databricks workspace
- create an Azure Databricks workspace
- create an Azure Databricks cluster
- create and run notebooks in Azure Databricks
- link and Azure Databricks workspace to an Azure Machine Learning workspace
2. Run Experiments and Train Models (20-25%)
Create models by using the Azure Machine Learning Designer
- create a training pipeline by using Azure Machine Learning designer
- ingest data in a designer pipeline
- use designer modules to define a pipeline data flow
- use custom code modules in designer
Run model training scripts
- create and run an experiment by using the Azure Machine Learning SDK
- configure run settings for a script
- consume data from a dataset in an experiment by using the Azure Machine Learning SDK
- run a training script on Azure Databricks compute
- run code to train a model in an Azure Databricks notebook
Generate metrics from an experiment run
- log metrics from an experiment run
- retrieve and view experiment outputs
- use logs to troubleshoot experiment run errors
- use MLflow to track experiments
- track experiments running in Azure Databricks
Use Automated Machine Learning to create optimal models
- use the Automated ML interface in Azure Machine Learning studio
- use Automated ML from the Azure Machine Learning SDK
- select pre-processing options
- select the algorithms to be searched
- define a primary metric
- get data for an Automated ML run
- retrieve the best model
Tune hyperparameters with Azure Machine Learning
- select a sampling method
- define the search space
- define the primary metric
- define early termination options
- find the model that has optimal hyperparameter values
3. Deploy and operationalize machine learning solutions (35-40%)
Select compute for model deployment
- consider security for deployed services
- evaluate compute options for deployment
Deploy a model as a service
- configure deployment settings
- deploy a registered model
- deploy a model trained in Azure Databricks to an Azure Machine Learning endpoint
- consume a deployed service
- troubleshoot deployment container issues
Manage models in Azure Machine Learning
- register a trained model
- monitor model usage
- monitor data drift
Create an Azure Machine Learning pipeline for batch inferencing
- configure a ParallelRunStep
- configure compute for a batch inferencing pipeline
- publish a batch inferencing pipeline
- run a batch inferencing pipeline and obtain outputs
- obtain outputs from a ParallelRunStep
Publish an Azure Machine Learning designer pipeline as a web service
- create a target compute resource
- configure an Inference pipeline
- consume a deployed endpoint
Implement pipelines by using the Azure Machine Learning SDK
- create a pipeline
- pass data between steps in a pipeline
- run a pipeline
- monitor pipeline runs
Apply ML Ops practices
- trigger an Azure Machine Learning pipeline from Azure DevOps
- automate model retraining based on new data additions or data changes
- refactor notebooks into scripts
- implement source control for scripts
4. Implement Responsible ML (5-10%)
Use model explainers to interpret models
- select a model interpreter
- generate feature importance data
Describe fairness considerations for models
- evaluate model fairness based on prediction disparity
- mitigate model unfairness
Describe privacy considerations for data
- describe principles of differential privacy
- specify acceptable levels of noise in data and the effects on privacy
Microsoft DP-100 Certification Sample Questions and Answers
To make you familiar with Designing and Implementing a Data Science Solution on Microsoft Azure (DP-100) certification exam structure, we have prepared this sample question set. We suggest you try our Sample Questions for Designing and Implementing a Data Science Solution on Azure DP-100 Certification to test your understanding of Microsoft DP-100process with the real Microsoft certification exam environment.
DP-100 Designing and Implementing a Data Science Solution on Microsoft Azure Sample Questions:-
01. You use the Two-Class Neural Network module in Azure Machine Learning Studio to build a binary classification model. You use the Tune Model Hyperparameters module to tune accuracy for the model.
You need to configure the Tune Model Hyperparameters module. Which two values should you use?
Each correct answer presents part of the solution. NOTE: Each correct selection is worth one point.
a) Number of hidden nodes
b) Learning Rate
c) The type of the normalizer
d) Number of learning iterations
e) Hidden layer specification
02. You are analyzing a dataset by using Azure Machine Learning Studio. You need to generate a statistical summary that contains the p-value and the unique count for each feature column.
Which two modules can you use?
Each correct answer presents a complete solution. NOTE: Each correct selection is worth one point.
a) Computer Linear Correlation
b) Export Count Table
c) Execute Python Script
d) Convert to Indicator Values
e) Summarize Data
03. You train a machine learning model. You must deploy the model as a real-time inference service for testing.
The service requires low CPU utilization and less than 48 MB of RAM. The compute target for the deployed service must initialize automatically while minimizing cost and administrative overhead.
Which compute target should you use?
a) Azure Container Instance (ACI)
b) attached Azure Databricks cluster
c) Azure Kubernetes Service (AKS) inference cluster
d) Azure Machine Learning compute cluster
04. You manually define security rules to isolate your Azure Machine Learning training processes within an Azure Virtual Network. You need to ensure that only required communications from the Internet are allowed.
Which three actions should you perform?
Each correct answer presents part of the solution. Choose the correct answers.
a) Define an inbound security rule.
b) Ensure the source service tag is set to AzureMachineLearning.
c) Ensure the source service tag is set to BatchNodeManagement.
d) Configure a source port range of 29876-29877.
e) Create an outbound security rule.
f) Configure a destination port range of 22.
05. You plan to use a Deep Learning Virtual Machine (DLVM) to train deep learning models using Compute Unified Device Architecture (CUDA) computations.
You need to configure the DLVM to support CUDA. What should you implement?
a) Solid State Drives (SSD)
b) Computer Processing Unit (CPU) speed increase by using overclocking
c) Graphic Processing Unit (GPU)
d) High Random Access Memory (RAM) configuration
e) Intel Software Extensions (Intel SGX) technology
06. You are building a regression model for estimating the number of calls during an event. You need to determine whether the feature values achieve the conditions to build a Poisson regression model.
Which two conditions must the feature set contain?
Each correct answer presents part of the solution. NOTE: Each correct selection is worth one point.
a) The label data must be a negative value.
b) The label data must be whole numbers.
c) The label data must be non-discrete.
d) The label data must be a positive value.
e) The label data can be positive or negative.
07. You use Azure Machine Learning designer to publish an inference pipeline as a web service. During deployment, you create a service principal (SP) and configure authentication.
You need to use the SP while consuming the endpoint. What should you do?
Choose the correct answer.
a) Use the client secret to retrieve an authentication token.
b) Use the regen_key method to regenerate the primary key.
c) Use the AciWebservice.deploy_configuration to set auth_enabled to True.
d) Use the get_keys method to retrieve authentication keys.
08. You retrain an existing model. You need to register the new version of a model while keeping the current version of the model in the registry. What should you do?
a) Register a model with a different name from the existing model and a custom property named version with the value 2.
b) Register the model with the same name as the existing model.
c) Save the new model in the default datastore with the same name as the existing model. Do not register the new model.
d) Delete the existing model and register the new one with the same name.
09. You plan to provision an Azure Machine Learning Basic edition workspace for a data science project. You need to identify the tasks you will be able to perform in the workspace.
Which three tasks will you be able to perform?
Each correct answer presents a complete solution. NOTE: Each correct selection is worth one point.
a) Create a Compute Instance and use it to run code in Jupyter notebooks.
b) Create an Azure Kubernetes Service (AKS) inference cluster.
c) Use the designer to train a model by dragging and dropping pre-defined modules.
d) Create a tabular dataset that supports versioning.
e) Use the Automated Machine Learning user interface to train a model.
10. You have access to structured data used in a previous machine learning project. The data is not registered in a dataset.
You need to use Azure Machine Learning SDK to use the data directly in a machine learning script. What should you do?
Choose the correct answer
a) Use Python to create a TabularDataset. Specify the path to the data.
b) Use the get_context method from the Run class to load the run from the remote environment.
c) Use the register method from the Dataset class to create a new dataset version.
d) Use the from_files method to create a dataset and register the dataset in your workspace.
Answers:-
Answer 1:- b, d
Answer 2:- b, e
Answer 3:- a
Answer 4:- a, c, d
Answer 5:- c
Answer 6:- b, d
Answer 7:- a
Answer 8:- a
Answer 9:- a, b, c, d
Answer 10:- a
Comments