The ETL (Extract, Transform, Load) process follows four main steps:
i) Connect and Collect: Connect to the data source/s and move data to local and crowdsource data storage.
ii) Data transformation using computing services such as HDInsight, Hadoop, Spark etc.
iii) Publish: To load data into Azure data lake storage, Azure SQL data warehouse, Azure SQL databases, Azure Cosmos DB, etc.
iv)Monitor: Azure Data Factory has built-in support for pipeline monitoring via Azure Monitor, API, PowerShell, Azure Monitor logs, and health panels on the Azure portal.
Build a unique job-winning data engineer resume with big data mini projects.