paint-brush
Building an ETL Pipeline to Load Data Incrementally from Office365 to S3 using ADF and Databricksby@yi
1,464 reads
1,464 reads

Building an ETL Pipeline to Load Data Incrementally from Office365 to S3 using ADF and Databricks

by Yi Ai4mNovember 19th, 2021
Read on Terminal Reader
Read this story w/o Javascript
tldt arrow

Too Long; Didn't Read

In this post, we will look at creating an Azure data factory with a pipeline that loads Office 365 event data incrementally based on change data capture (CDC) information in the source of Change Data Feed(CDF) of a Delta lake table to an AWS S3 bucket. What we’ll cover: Create an ADF Pipeline that loads Calendar events from Offfice365 to a Blob container. Run a Databricks Notebook with the activity in the ADF pipeline, transform extracted Calendar event and merge to a Delta Lake table.

Companies Mentioned

Mention Thumbnail
Mention Thumbnail
featured image - Building an ETL Pipeline to Load Data Incrementally from Office365 to S3 using ADF and Databricks
Yi Ai HackerNoon profile picture
Yi Ai

Yi Ai

@yi

L O A D I N G
. . . comments & more!

About Author

Yi Ai HackerNoon profile picture
Yi Ai@yi

TOPICS

THIS ARTICLE WAS FEATURED IN...

Permanent on Arweave
Read on Terminal Reader
Read this story in a terminal
 Terminal
Read this story w/o Javascript
Read this story w/o Javascript
 Lite