Agile Metrics That Matter - Prachi - Maini

Download as pptx, pdf, or txt
Download as pptx, pdf, or txt
You are on page 1of 25

Agile Metrics that Matter

Prachi Maini
Manager, QA Engineering
Morningstar, Inc.

©2015 Morningstar, Inc. All rights reserved.


Executive Summary

• Define Metrics that


The can be used by
Agile teams and
Concept Team management

The • Reduced costs


• Increased team
Opportunity satisfaction

• Auto Generate
The using exposed APIs
provided by various
Potential tools

2
Independent investment research &
management firm headquartered in
Chicago.
Consumer report for securities.
Agile Squads (5-7dev, 1-2 QA, product
owner, scrum master, designer).
Two week sprints.
Toolset
• QAC for test case management
• Selenium for functional automation
• ReadyAPI for webservices
• Webload for Performance

3
Why Do We Need Metrics

Drive strategy and direction.


Provide measurable data and trend to the
project team and management.
Ensure that the project remains on track.
Quantify risk and process improvements.
Ensure customer satisfaction with the
deployed product.
Assist with resource/budget estimation and
forecasting.

4
Effective Metrics

Metrics should be clearly defined so the


team or organization can benchmark its
success.
Secure buy-in from management and
employees.
Have clearly defined data and collection
process.
Are measurable and shared.
Can be automatically generated and
scheduled.
5
Agile QA Dashboard
Metrics are role agnostic.
Focus on trends rather than absolute values.
Used by teams to perform introspection on
their own performance and feed into release
planning
Core Agile metrics should not be used to
compare different teams or focus on
underperforming teams.

6
Project Progress
• Burndown Chart
– Graphical representation of work remaining
vs time.
• Committed vs Completed
– The percentage of points completed by the
squad as a percentage of the committed
points for the sprint
• Tech Category
– This helps identify how an agile team is
spending its time. The possible values for
tech category can be client customization,
new product development, operations or
maintenance. 7
Committed vs Completed
200
Committed vs Sprint
Completed Sprint 74 Sprint 75 Sprint 76 77 Sprint 78 150

Committed 148 105 98 53 154 100


Completed 74 87 75 51 125
50

0
Sprint Sprint Sprint Sprint Sprint
74 75 76 77 78
Committed Completed

Work done by Tech Category


120
Sprint 74 Sprint 75 Sprint 76 Sprint 77 Sprint 78 100
Tech Category 80
60
Client Customization 0 0 0 6 16
40
New Product Development 157 171 91 144 109 20
Maintenance 57 54 52 46 59 0
Sprint 74 Sprint 75 Sprint 76 Sprint 77 Sprint 78
Other 23 10 0 0 0
Other Maintenance

8
New Product Development Client Customization
What To Watch For
The team finishes early sprint after sprint
because they are not committing enough
points
The team is not meeting its commitment
because they are overcommitting each sprint.
Burndown is steep rather than gradual
because work is not broken down into granular
units.
Scope is often added or changed mid-sprint.

9
Velocity
• Velocity
– Points of work completed by an agile team within
a given sprint
• Adjusted Velocity
– Points of work completed by an agile team
accounting for holidays, team absence etc.
– Calculated as velocity / available man days.
– Running average of last three sprints is reported.

10
Sprint Sprint Adjusted Velocity
Capacity Sprint 74 75 76 Sprint 77 Sprint 78
150
Team Size 8 8 8 8 8
Available Days 80 80 80 80 80 100
Unavailable Days 10 12 11 5 0
50
Net Days (Capacity) 70 68 69 75 80
Velocity 0
Total Points Completed 73 87 75 51 125 Sprint 74 Sprint 75 Sprint 76 Sprint 77 Sprint 78
Adjusted Velocity 83 102 87 54 125 Adjusted Velocity Avg Velocity (Last 3 Sprints)
Avg Velocity (Last 3
Sprints) 88 90 91 81 89

11
What To Watch For
 An erratic average velocity over a period of
time requires revisiting the team’s estimation
practices.
Are there unforeseen challenges not
accounted for when estimating the work
DO NOT
Use velocity to compare two different teams
since the level of work estimation is different
from team to team
Use velocity to identify lower performing teams.

12
Quality of Code
• First Pass Rate
– Used for measuring the amount of rework in
the process
– Defined as number of test cases passed on
first execution.
FPR = Passed\Total on First Execution
– For stories that deal with the development of
new APIs or Features
– For stories that deal with addendums to APIs
or Features FPR should include regression.

13
Story Total TCs Pass Fail FPR
Navy
PHX-10112 2 1 1 0.5
PHX-10411 8 6 2 0.75
PHX- 10382 15 8 7 0.8
PHX- 7703 10 6 4 6
PHX - 10336 1 1 0 1
34 21 13 0.62

14
What To Watch For
Lower first pass rates indicate that Agile tools
like desks checks , unit testing are not used
sufficiently.
Lower first pass rate could indicate lack of
understanding of requirements.
Higher first pass rate combined with high
defect rate in production could indicate lack of
proper QA.

15
Bug Dashboard
– Net Open Bugs / Created vs Resolved
• This gives a view of the team flow rate. Are we
creating more technical debt and defects than
what the team can resolve.
– Functional Vs Regression Bugs Trend
• This helps identify the defects found in new
development vs regression.
– Defects Detected in
• This helps identify the environment in which the
defect is detected. (QA, Staging, UAT, Production)
• For defects detected in environment higher than
QA, an RCA is needed.

16
Net Open Bugs
Defects : Net Open Defects 470
465
460
Sprint 74 Sprint 75 Sprint 76 Sprint 77 Sprint 78 455
450
445
Bugs Opened 68 36 33 41 17 440
435
Bugs Closed 30 30 16 38 15 430
Net Open 425
Bugs 438 444 461 464 466 420
Sprint 74 Sprint 75 Sprint 76 Sprint 77 Sprint 78

Net Open Bugs

Defects : Regression vs
Feature 80

70
Sprint 74 Sprint 75 Sprint 76 Sprint 77 Sprint 78 68
60
Regression 16 24 16 38 40
50
Feature 68 36 33 41 17
40 41 40
38 Regression
36
33 Feature
30
24
20
16 16 17
10

0
Sprint 74 Sprint 75 Sprint 76 Sprint 77 Sprint 78

17
35
30 30 30
Defects Found By Environment 25 25
28
26
QA
Sprint 74 Sprint 75 Sprint 76 Sprint 77 Sprint 78 20
Stg
15
25 30 28 30 32 UAT
QA 10
Staging 5 4 4 4 3 8 Prod
5 5 4 4 4
UAT 1 3 1 3 1 2 3 2 3 3
0 1 1 1 1 1
Prod 2 1 2 1 4 Sprint 74 Sprint 75 Sprint 76 Sprint 77 Sprint 78…

Defects : Root Cause (Non 25


21 21 22
20
Production)
Sprint 74 Sprint 75 Sprint 76 Sprint 77 Sprint 78 15 15
19
Code
Config
Code 15 21 21 22 19 10 9 8 Requirements
7 6
Config 5 4 4 6 3 5 5 4 4 4 5 Hardware
2 3 2 3 3
Requirements 1 3 1 3 1 0 1 1 1 1 1 Data
Hardware 2 1 2 1 8 Sprint 74 Sprint 75 Sprint 76 Sprint 77 Sprint 78
Data 2 9 7 4 5

25
22
Defects : Root Cause (Production)
20
21 21
19

15 15 QA oversight
Sprint 74 Sprint 75 Sprint 76 Sprint 77 Sprint 78
Environment
10
QA oversight 15 21 21 22 19 Requirements
Environment 5 4 4 6 3 6 6 Existing Issue
5 5 5
Requirements 1 3 1 3 1 4 4 4 4 Data
3 3 3
Existing Issue 2 1 2 1 8 0
1 1 1
Data 4 4 5 4 6 Sprint 74 Sprint 75 Sprint 76 Sprint 77 Sprint 78

18
What To Watch For
An increase in regression bug count indicates
the impact of code refactoring.
An increase bug count in non-QA environment
due to environment differences requires
revisiting the environment strategy.
 An increase bug count in non-QA environment
due to QA oversight requires revisiting the
testing strategy.

19
Automation
• Number of automated test cases
– Percentage of automated test cases as part of
total automation candidates
– Percentage of automated test cases as part of
total test cases
– Can be reported separately for API ,
Functional , Migration testing

• Defects Found by Automation


– Number of defects found via automation and
manual testing

20
Automation Progress
V2 Functional Test Cases Sprint 74 Sprint 75 Sprint 76 Sprint 77 Sprint 78
# of Total Test Cases 2444 2611 2684 2782 2822
# of Automatable Test Cases 1541 1642 1690 1755 1774
# of Test Cases Automated 1138 1234 1267 1339 1351
% Coverage of Automatable Test Cases 73.85% 75.15% 74.97% 76.30% 76.16%
% Coverage of Total Test Cases 47% 47.26% 47% 48.13% 48%

3000 90.00%
80.00% # of Total Test Cases
2500
70.00%
2000 60.00% # of Automatable Test Cases
50.00%
1500
40.00% # of Test Cases Automated

1000 30.00%
% Coverage of Automatable
20.00% Test Cases
500
10.00%
% Coverage of Total Test
0 0.00% Cases
Sprint Sprint Sprint Sprint Sprint
74 75 76 77 78

AEM Regression Testing Window


140
120 Execution Time
100 (Average - min)

80 Analysis Time
(Average - min)
60
Total Time (Average -
40 Min)
20
0
Jan-16 Feb-16 Mar-16 Apr-16
21
What to watch for

A decrease in automation coverage could


indicate that a lot of automation capacity is
being spent in script maintenance.
The coverage helps in identifying the
percentage of application that can be
effectively monitored and regressed on a
recurring basis

22
Team Sentiments
• Survey questions distributed to Agile
Teams
– Team members are anonymously asked to
rate on a scale of 1 to 10 on questions
pertinent to the project. Example of question
include but are not limited to
• Understanding of the vision of the project
• Quality of the user stories
• Collaboration between team members
– Responses are tabulated and shared with the
team
– Trends are noted over time to identify teams
alignment with the company and project
vision and general satisfaction with the project 23
24
Thank You!!

• Questions
• Prachi Maini
[email protected]
[email protected]
• https://www.linkedin.com/in/prachimaini
• Phone: 630-818-6472

25

You might also like