Azure User Stories - Updated
Azure User Stories - Updated
User Story 1: Create an ADF Pipeline to read and write files in Azure Data Lake Gen 2 using
Azure Data Factory
Description:
Azure Data Lake Gen 2(ADLS) is a storage service provided by Microsoft Azure which allows you
to save data in multiple formats.
Create an input and output container with the naming convention input_modakid and
output_modakid (example input_vkg92599)
Upload the files attached in this story to the input container of the datalake
Create a pipeline in Azure Data Factory(ADF) to migrate the files from input to output
container.
Acceptance Criteria:
Story Points: 1
---------------------------------------------------------------------------------------------------------------------
User Story 2: Create an ADF pipeline to migrate the data from a file in ADLS Gen1 to Azure SQL
Database table
Description:
Access the ADLS Gen1 storage and upload file(csv/excel/parquet) into the datalake
Create a table in Azure SQL database which maps with the schema of file uploaded in
datalake
Create a pipeline using the Copy data activity available on the ADF Landing page
Acceptance Criteria:
Validate the migration of data using with row counts in the raw file and SQL Table
Record (Screenshots) and document the steps
verify the story with your team lead
Story Points: 1
---------------------------------------------------------------------------------------------------------------------
User Story 3: Create an ADF pipeline to copy data from ADLS Gen1/Gen2 to Azure SQL
Database
Description:
Upload a file attached ‘ch’ into Gen1/Gen2
Create a table in Azure SQL database which maps with the schema after transformations
(Add a prefix as_ to all the column names while creating the table)
Create a pipeline to copy data from a file in ADLS Gen1/Gen2 to Azure SQL Database
Acceptance Criteria:
Story Points: 2
---------------------------------------------------------------------------------------------------------------------
User Story 4: Copy multiple files with same type of data(schema) from ADLS Gen 2 to Azure SQL
Description:
Upload 4 files into a folder in ADLS Gen2 storage (All the files must contain same type of
date i.e. same schema)
Create a table in Azure SQL database which maps with the schema of files uploaded in
datalake
Create a pipeline to copy data from all the files in the datalake to a single table in Azure
SQL Database
Acceptance Criteria:
Calculate the total rows in Source files and validate the count with the rows in Azure
SQL Database
Record (Screenshots) and document the steps
Verify the story with your team lead
Story Points: 2
--------------------------------------------------------------------------------------------------------------------
As a Data Analyst,
I want Azure Data Factory to orchestrate a scheduled ETL pipeline that moves new patient
vaccination records from ADLS Gen2 to Azure SQL,
So that I can analyze real-time vaccination data and generate up-to-date reports.
Acceptance Criteria:
Story Points: 3
---------------------------------------------------------------------------------------------------------------------
Story Points: 1
---------------------------------------------------------------------------------------------------------------------
User Story 7: Analysis in Databricks
Description:
Load the file to DBFS in Databricks
Read the file from DBFS
Analysis in Databricks
Count vaccinations by month.
Identify top locations by vaccination count.
------------------------------------------------------------------------------------------------------------------------------