Azure Data Factory

Nitish Kumar
2 min readJul 30, 2022

Azure Data Factory is a cloud-based data integration service that allows you to orchestrate and automate data movement and transformation using cloud-based data-driven workflows. It provides a cost-effective solution whenever a dependable code-free ETL tool on the cloud with numerous integrations is required.

Azure Data Factory consists of a variety of components like:-

  • A pipeline is a logical grouping of activities that perform a task. It allows you to manage the activity as a set instead of each one individually. Ingesting data from the storage blob, querying the SQL Database, and a single pipeline can perform many other activities.
  • Activities in a pipeline represent a unit of work. It defines the actions on your data, like copying data from SQL Server to Azure Blob Storage.
  • Linked services are connection strings that allow you to connect to the data stores.
  • Datasets represent the data structures within the data stores that point to the data you want to use as inputs and outputs in your activities.
  • Triggers are a method for executing a pipeline run. Triggers specify when the execution of a pipeline should begin. There are three types of triggers that Data Factory currently supports:

=> Schedule trigger

=> Tumbling window trigger

=> Event-based trigger

Thank you!

Sign up to discover human stories that deepen your understanding of the world.

Free

Distraction-free reading. No ads.

Organize your knowledge with lists and highlights.

Tell your story. Find your audience.

Membership

Read member-only stories

Support writers you read most

Earn money for your writing

Listen to audio narrations

Read offline with the Medium app

Nitish Kumar
Nitish Kumar

Written by Nitish Kumar

Data engineer / Python / NodeJS

No responses yet

Write a response