0% found this document useful (0 votes)
8 views4 pages

Azure User Stories - Updated

Uploaded by

melvanogondra
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
8 views4 pages

Azure User Stories - Updated

Uploaded by

melvanogondra
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 4

-----------------------------------------------------------------------------------------------------------------------------

User Story 1: Create an ADF Pipeline to read and write files in Azure Data Lake Gen 2 using
Azure Data Factory
Description:

Azure Data Lake Gen 2(ADLS) is a storage service provided by Microsoft Azure which allows you
to save data in multiple formats.

 Create an input and output container with the naming convention input_modakid and
output_modakid (example input_vkg92599)

 Upload the files attached in this story to the input container of the datalake
 Create a pipeline in Azure Data Factory(ADF) to migrate the files from input to output
container.

Acceptance Criteria:

 Validate the migration of files using UI


 open the csv file to verify if the data was migrated in the same format
 Record (Screenshots) and document the steps
 verify the story with your team lead

Story Points: 1
---------------------------------------------------------------------------------------------------------------------
User Story 2: Create an ADF pipeline to migrate the data from a file in ADLS Gen1 to Azure SQL
Database table
Description:
 Access the ADLS Gen1 storage and upload file(csv/excel/parquet) into the datalake
 Create a table in Azure SQL database which maps with the schema of file uploaded in
datalake
 Create a pipeline using the Copy data activity available on the ADF Landing page
Acceptance Criteria:

 Validate the migration of data using with row counts in the raw file and SQL Table
 Record (Screenshots) and document the steps
 verify the story with your team lead

Story Points: 1
---------------------------------------------------------------------------------------------------------------------

User Story 3: Create an ADF pipeline to copy data from ADLS Gen1/Gen2 to Azure SQL
Database
Description:
 Upload a file attached ‘ch’ into Gen1/Gen2
 Create a table in Azure SQL database which maps with the schema after transformations
(Add a prefix as_ to all the column names while creating the table)
 Create a pipeline to copy data from a file in ADLS Gen1/Gen2 to Azure SQL Database
Acceptance Criteria:

o Validate the migration of data


o Record (Screenshots) and Document the steps
o verify the story with your team lead

Story Points: 2
---------------------------------------------------------------------------------------------------------------------
User Story 4: Copy multiple files with same type of data(schema) from ADLS Gen 2 to Azure SQL
Description:
 Upload 4 files into a folder in ADLS Gen2 storage (All the files must contain same type of
date i.e. same schema)
 Create a table in Azure SQL database which maps with the schema of files uploaded in
datalake
 Create a pipeline to copy data from all the files in the datalake to a single table in Azure
SQL Database
Acceptance Criteria:

 Calculate the total rows in Source files and validate the count with the rows in Azure
SQL Database
 Record (Screenshots) and document the steps
 Verify the story with your team lead
Story Points: 2
--------------------------------------------------------------------------------------------------------------------

User Story 5: Scheduled ETL Pipeline for Real-Time Analytics

As a Data Analyst,

I want Azure Data Factory to orchestrate a scheduled ETL pipeline that moves new patient
vaccination records from ADLS Gen2 to Azure SQL,

So that I can analyze real-time vaccination data and generate up-to-date reports.

Acceptance Criteria:

1. ADF schedules and triggers the pipeline hourly.


2. Data is extracted from ADLS Gen2, loaded into Azure SQL.

Story Points: 3
---------------------------------------------------------------------------------------------------------------------

User Story 6: Move data from Azure SQL Database to Gen1/Gen2


Description:

 Create a table in Azure SQL Database with duplicate data.


 Create a stored procedure to get distinct rows from the table.
 Create a pipeline to migrate data from source to destination.
Acceptance Criteria:

 Record (Screenshots) and Document the steps


 verify the story with your team lead

Story Points: 1
---------------------------------------------------------------------------------------------------------------------
User Story 7: Analysis in Databricks
Description:
 Load the file to DBFS in Databricks
 Read the file from DBFS
 Analysis in Databricks
 Count vaccinations by month.
 Identify top locations by vaccination count.

------------------------------------------------------------------------------------------------------------------------------

Sample Data for User story 5 and 7


VaccineTy DoseNum VaccinationDa
PatientID Name Age Gender pe ber te Location
New York,
1001 John Smith 28 Male Pfizer 1 2023-01-15 USA
1002 Jane Doe 35 Female Moderna 1 2023-02-10 London, UK
Alice Sydney,
1003 Johnson 42 Female Pfizer 2 2023-03-05 Australia
Bob AstraZenec Toronto,
1004 Williams 50 Male a 1 2023-01-20 Canada
Maria Madrid,
1005 Gonzalez 30 Female Moderna 2 2023-04-18 Spain
Ahmed
1006 Khan 25 Male Pfizer 1 2023-02-25 Dubai, UAE
1007 Emily Chen 29 Female Pfizer 2 2023-05-02 Singapore
David New York,
1008 Brown 55 Male Moderna 1 2023-03-22 USA
AstraZenec Sydney,
1009 Sarah Lee 38 Female a 2 2023-04-15 Australia

You might also like