Skip to content

ds-fau-ck/Aero-Data-Pipeline-in-Azure

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

End-to-End Airline Data Ingestion Pipeline in Azure

Introduction

In the Azure environment, a dedicated Data Factory (ADF) namespace is used for development. During development, data is read from the storage account named development, while a separate storage account is used for production.

Architecture

Architecture!

The steps for setting up ADF in the development environment are as follows:

  1. Linked Service Creation: First, a linked service is set up for the storage account. This step involves specifying the storage account name, authentication details, and connection string for the storage account.

  2. Dataset Creation: The next step is to create datasets for both the source and target locations. A dataset is configured to point to the exact location of the file being read, as well as the target storage account where the data will be written.

  3. Pipeline Creation: Finally, a pipeline is created in ADF to orchestrate the data flow between the source and target storage accounts.

This outlines the development process for ADF in the Azure environment.

Technology Used

Azure Cloud Platform:

  • Azure Data Lake Storage (ADLS)
  • Azure DevOPS
  • Azure Data Factory

Data used for this Project

  1. Airport
  2. Flight

Releases

No releases published

Packages

No packages published