Azure Data Factory

Nitish Kumar
2 min readJul 30, 2022

--

Azure Data Factory is a cloud-based data integration service that allows you to orchestrate and automate data movement and transformation using cloud-based data-driven workflows. It provides a cost-effective solution whenever a dependable code-free ETL tool on the cloud with numerous integrations is required.

Azure Data Factory consists of a variety of components like:-

  • A pipeline is a logical grouping of activities that perform a task. It allows you to manage the activity as a set instead of each one individually. Ingesting data from the storage blob, querying the SQL Database, and a single pipeline can perform many other activities.
  • Activities in a pipeline represent a unit of work. It defines the actions on your data, like copying data from SQL Server to Azure Blob Storage.
  • Linked services are connection strings that allow you to connect to the data stores.
  • Datasets represent the data structures within the data stores that point to the data you want to use as inputs and outputs in your activities.
  • Triggers are a method for executing a pipeline run. Triggers specify when the execution of a pipeline should begin. There are three types of triggers that Data Factory currently supports:

=> Schedule trigger

=> Tumbling window trigger

=> Event-based trigger

Thank you!

--

--

Nitish Kumar
Nitish Kumar

Written by Nitish Kumar

Data engineer / Python / NodeJS

No responses yet