Course Overview

Note: Exam DP-203 is replacing exams DP-200 and DP-201. DP-200 and DP-201 will retire on June 30, 2021.

The DP-203 Data Engineering on Microsoft Azure certification training course from CloudThat offers candidates proper training and relevant study material to prepare and successfully clear the DP-203 exam.

After completing this course, students will be able to:

  • Design and implement data storage
  • Design and develop data processing
  • Design and implement data security
  • Monitor and optimize data storage and data processing

Upcoming Batches

India Online Enroll
Start Date End Date

03-10-2022

06-10-2022

03-10-2022

06-10-2022

07-11-2022

10-11-2022

07-11-2022

10-11-2022

08-10-2022

16-10-2022

10-10-2022

13-10-2022

10-10-2022

13-10-2022

14-11-2022

17-11-2022

14-11-2022

17-11-2022

17-10-2022

20-10-2022

17-10-2022

20-10-2022

24-10-2022

27-10-2022

24-10-2022

27-10-2022

26-09-2022

29-09-2022

26-09-2022

29-09-2022

31-10-2022

03-11-2022

31-10-2022

03-11-2022

Key Features

  • Our training modules have 50% - 60% hands-on lab sessions to encourage Thinking-Based Learning (TBL)
  • Interactive-rich virtual and face-to-face classroom teaching to inculcate Problem-Based Learning (PBL)
  • Microsoft certified instructor-led training and mentoring sessions to develop Competency-Based Learning (CBL)
  • Well-structured use-cases to simulate challenges encountered in a Real-World environment
  • Integrated teaching assistance and support through experts designed Learning Management System (LMS) and ExamReady platform
  • Being a Microsoft Learning Partner provides us with the edge over competition

Who Should Attend

  • Azure Data Engineers who integrate, transform, and consolidate data from various structured and unstructured data systems into structures that are suitable for building analytics solutions

Prerequisites

The prerequisites of DP-203 exam include:

  • A candidate must have solid knowledge of data processing languages, such as SQL, Python, or Scala, and they need to understand parallel processing and data architecture patterns.
  • Candidate should have subject matter expertise integrating, transforming, and consolidating data from various structured and unstructured data systems into structures that are suitable for building analytics solutions.

Course Outline Download Course Outline

  • Design an Azure Data Lake solution
  • Recommend file types for storage
  • Recommend file types for analytical queries
  • Design for efficient querying
  • Design for data pruning
  • Design a folder structure that represents the levels of data transformation
  • Design a distribution strategy
  • Design a data archiving solution
  • Design a partition strategy
  • Design a partition strategy for files
  • Design a partition strategy for analytical workloads
  • Design a partition strategy for efficiency/performance
  • Design a partition strategy for Azure Synapse Analytics
  • Identify when partitioning is needed in Azure Data Lake Storage Gen2

Design the serving layer

  • Design star schemas
  • Design slowly changing dimensions
  • Design a dimensional hierarchy
  • Design a solution for temporal data
  • Design for incremental loading
  • Design analytical stores
  • Design metastores in Azure Synapse Analytics and Azure Databricks

Implement physical data storage structures

  • Implement compression
  • Implement partitioning
  • Implement sharding
  • Implement different table geometries with Azure Synapse Analytics pools
  • Implement data redundancy
  • Implement distributions
  • Implement data archiving

Implement logical data structures

  • Build a temporal data solution
  • Build a slowly changing dimension
  • Build a logical folder structure
  • Build external tables
  • Implement file and folder structures for efficient querying and data pruning

Implement the serving layer

  • Deliver data in a relational star schema
  • Deliver data in Parquet files
  • Maintain metadata
  • Implement a dimensional hierarchy

  • Transform data by using Apache Spark
  • Transform data by using Transact-SQL
  • Transform data by using Data Factory
  • Transform data by using Azure Synapse Pipelines
  • Transform data by using Stream Analytics
  • Cleanse data
  • Split data
  • Shred JSON
  • Encode and decode data
  • Configure error handling for the transformation
  • Normalize and denormalize values
  • Transform data by using Scala
  • Perform data exploratory analysis

Design and develop a batch processing solution

  • Develop batch processing solutions by using Data Factory, Data Lake, Spark, Azure Synapse Pipelines, PolyBase, and Azure Databricks
  • Create data pipelines
  • Design and implement incremental data loads
  • Design and develop slowly changing dimensions
  • Handle security and compliance requirements
  • Scale resources
  • Configure the batch size
  • Design and create tests for data pipelines
  • Integrate Jupyter/IPython notebooks into a data pipeline
  • Handle duplicate data
  • Handle missing data
  • Handle late-arriving data
  • Upsert data
  • Regress to a previous state
  • Design and configure exception handling
  • Configure batch retention
  • Design a batch processing solution
  • Debug Spark jobs by using the Spark UI

Design and develop a stream processing solution

  • Develop a stream processing solution by using Stream Analytics, Azure Databricks, and Azure Event Hubs
  • Process data by using Spark structured streaming
  • Monitor for performance and functional regressions
  • Design and create windowed aggregates
  • Handle schema drift
  • Process time series data
  • Process across partitions
  • Process within one partition
  • Configure checkpoints/watermarking during processing
  • Scale resources
  • Design and create tests for data pipelines
  • Optimize pipelines for analytical or transactional purposes
  • Handle interruptions
  • Design and configure exception handling
  • Upsert data
  • Replay archived stream data
  • Design a stream processing solution

Manage batches and pipelines

  • Trigger batches
  • Handle failed batch loads
  • Validate batch loads
  • Manage data pipelines in Data Factory/Synapse Pipelines
  • Schedule data pipelines in Data Factory/Synapse Pipelines
  • Implement version control for pipeline artifacts
  • Manage Spark jobs in a pipeline

  • Design data encryption for data at rest and in transit
  • Design a data auditing strategy
  • Design a data masking strategy
  • Design for data privacy
  • Design a data retention policy
  • Design to purge data based on business requirements
  • Design Azure role-based access control (Azure RBAC) and POSIX-like Access Control List (ACL) for Data Lake Storage Gen2
  • Design row-level and column-level security

Implement data security

  • Implement data masking
  • Encrypt data at rest and in motion
  • Implement row-level and column-level security
  • Implement Azure RBAC
  • Implement POSIX-like ACLs for Data Lake Storage Gen2
  • Implement a data retention policy
  • Implement a data auditing strategy
  • Manage identities, keys, and secrets across different data platform technologies
  • Implement secure endpoints (private and public)
  • Implement resource tokens in Azure Databricks
  • Load a DataFrame with sensitive information
  • Write encrypted data to tables or Parquet files
  • Manage sensitive information

  • Implement logging used by Azure Monitor
  • Configure monitoring services
  • Measure performance of data movement
  • Monitor and update statistics about data across a system
  • Monitor data pipeline performance
  • Measure query performance
  • Monitor cluster performance
  • Understand custom logging options
  • Schedule and monitor pipeline tests
  • Interpret Azure Monitor metrics and logs
  • Interpret a Spark directed acyclic graph (DAG)

Optimize and troubleshoot data storage and data processing

  • Compact small files
  • Rewrite user-defined functions (UDFs)
  • Handle skew in data
  • Handle data spill
  • Tune shuffle partitions
  • Find shuffling in a pipeline
  • Optimize resource management
  • Tune queries by using indexers
  • Tune queries by using cache
  • Optimize pipelines for analytical or transactional purposes
  • Optimize pipeline for descriptive versus analytical workloads
  • Troubleshoot a failed spark job
  • Troubleshoot a failed pipeline run

Certification

    • By earning DP-203 certification, you can become Microsoft Certified Azure Data Engineer
    • Demonstrate abilities to Design and implement data storage, data processing and data security features
    • On successful completion of DP-203: Data Engineering on Microsoft Azure training aspirants receive a Course Completion Certificate from us
    • By successfully clearing the DP-203 exams, aspirants earn Microsoft Certification

Course Fee

Select Course date

Add to Wishlist

Course ID: 8135

Course Price at

₹ 29900 + 18% GST

Enroll Now

Reviews

A

Asif Ali

Excellent training sessions provided by CloudThat. I have attended a few webinars on Microsoft Azure and the trainers are really knowledgeable with good real time experience on Azure Cloud. The materials and the test prep kit along with the interactive training sessions really helps in clearing the certification exams. I would recommend everyone who is looking to make a career in cloud domain to register for the trainings provided by CloudThat.

J

Jawed Akhtar

I had attend the Microsoft Azure training today.it was so good and nice to explain very clearly and it was really helpful for my upcoming professional careers.

R

Remya Ravi

Great and valuable training session. Thank you.