Sample Process Performance Report
Sample Process Performance Report
FOR:
XYZ Company
Global Development Center
CONTENT:
Process Performance Baseline Report
October to December, 2007
R-OPP-PPBR-Q22009
Version 1.2
11-Aug-2009
User of this document is responsible to use only the current version available in organizational database.
This document is for intended recipients only. No part of this document shall be reproduced, stored in retrieval system or transmitted in any form or by any means –
recording, photocopying, electronic and mechanical without prior permission of XYZ Company.
XYZ Company QMG R-OPP-PPBR-Q22009
Process Performance Baseline Report
April to June, 2009
v1.2
Document History
Version Date Author Description of Change Reviewed By Approved By
Distribution List
Designations
Designated PMs, Principals, VPs, QA team and Metrics Council members of other locations.
Mode of distribution will be through the HQ. This document is controlled through a configuration
management tool by Corporate Quality Manager and its revisions are identified by its version number.
Page 2 of 15
XYZ Company QMG R-OPP-PPBR-Q22009
Process Performance Baseline Report
April to June, 2009
v1.2
Reference Documents
Document ID
Page 3 of 15
XYZ Company QMG R-OPP-PPBR-Q22009
Process Performance Baseline Report
April to June, 2009
v1.2
TABLE OF CONTENTS
1 EXECUTIVE SUMMARY................................................................................................................................5
2 INTRODUCTION............................................................................................................................................10
2.1 Objectives...........................................................................................................................................................10
2.2 Scope..................................................................................................................................................................10
3 ACRONYMS AND DEFINITIONS...............................................................................................................10
4 ORGANIZATIONAL GOALS.......................................................................................................................12
4.1 Maintenance Projects:......................................................................................................................................12
4.2 Enhancement & Development Projects:..........................................................................................................12
4.3 Support Projects:...............................................................................................................................................13
4.4 QA (Testing) Projects:......................................................................................................................................13
5 MAINTENANCE PROJECTS.......................................................................................................................14
6 ENHANCEMENT AND DEVELOPMENT PROJECTS............................................................................14
7 SUPPORT PROJECTS...................................................................................................................................14
8 QA (TESTING) PROJECTS...........................................................................................................................14
9 RECOMMENDATIONS.................................................................................................................................14
10 APPENDIX - A: LIST OF DEFAULTING PROJECTS.............................................................................15
11 APPENDIX - B: RULES FOR DEFINING CONTROL LIMITS & GRAPHS PLOTTING..................15
Page 4 of 15
XYZ Company QMG R-OPP-PPBR-Q22009
Process Performance Baseline Report
April to June, 2009
v1.2
1 EXECUTIVE SUMMARY
This report provides analysis of the metrics data, consolidated for the projects being implemented
in GDCs at A & B locations. These projects belong to all business units. The data collected is for
the period from 1st April to 30th June, 2009. The highlights of the report are described below:
Highlights:
A total of 136 (80+43+7+6) projects are considered for analysis in this Process Performance
Baseline Report (PPBR). Break-up of the 136 projects are:
o All 80 out of 80 Maintenance projects.
o 43 [40 Enhancement + 3 Development] out of 54 projects.
o 7 out of 10 Support projects.
o 6 out of 8 Manual Testing Projects.
Maintenance and Support metrics are collected from the project trackers for each calendar
month.
Development and Enhancement metrics are collected on the basis of devolvement life cycle
phases (Requirement, Design, Coding, Testing and overall).
Testing metrics are collected from the project trackers for each release
The data distribution is checked for each metrics based on percentile method.
The Outliers in this PPBR are identified using Box Plots.
The Defect Rates reported in the two ‘Quantitative Performance 2008’ tables below are
much less than the actual as 60%% of projects have not logged defect data or have reported
zero defects.
Page 5 of 15
XYZ Company QMG R-OPP-PPBR-Q22009
Process Performance Baseline Report
April to June, 2009
v1.2
For Maintenance Projects, the trends of Performance Mean of various metrics are:
Page 6 of 15
XYZ Company QMG R-OPP-PPBR-Q22009
Process Performance Baseline Report
April to June, 2009
v1.2
Page 7 of 15
XYZ Company QMG R-OPP-PPBR-Q22009
Process Performance Baseline Report
April to June, 2009
v1.2
For Support Projects, the trends of Performance Mean of various metrics are:
Page 8 of 15
XYZ Company QMG R-OPP-PPBR-Q22009
Process Performance Baseline Report
April to June, 2009
v1.2
For QA (Testing) Projects, the trends of Performance Mean of various metrics are:
Conclusions:
1. If we refer to SPIN or ISBSG benchmarks and compare them with our data trends, one can
easily infer that most of the projects still log either “good looking data” in their trackers or take
it as a mere formality and assign a junior team member to ’fill’ the tracker once in a fortnight
to avoid escalations from PPQAs. Unless this attitude changes, we will not be able to use
measurement data for meaningful analysis leading to ‘data based decision making”.
2. Not all QA and Support projects have provided data for analysis and goal setting. However,
analysis of the same is done on the available data points.
3. Goal setting for most of the metrics covered in the PPBR have been done based on
benchmarks or the decision taken in Measurement Council. This is because the data being
logged in the Project Tracker is still to reflect the true status.
4. Projects not reporting their data for statistical analysis are listed in Appendix-A at the end of
the report.
5. Measurement Council decides to continue with the performance goals already
published in September 08, as revising the goals every quarter may confuse the
project teams.
If we compare this set of performance data with International benchmarks mentioned above in
the last column, it becomes evident that the data being reported by our projects does not
represent the true status of activities on the floor. However, more projects have started reporting
data which is a welcome change.
Next Steps:
QA have recently suggested to make AMCs or PMs (those who own responsibility of projects’
revenue) accountable for the quality of data in the project trackers. This suggestion, if
implemented, will certainly improve the current situation. The council decided to examine the
effectiveness of this suggestion while compiling the next PPBR.
2 INTRODUCTION
© XYZ Company-Restricted Circulation
Page 9 of 15
XYZ Company QMG R-OPP-PPBR-Q22009
Process Performance Baseline Report
April to June, 2009
v1.2
This Process Performance Baseline Report (PPBR) establishes Mean Performance values,
data variation patterns and statistically driven control limits (UCL and LCL), which in turn
act as organization benchmarks for various metrics. This also helps in comparing the current
performance values with those in the recent past. These values are then used by projects to set /
update product and process performance goals once in a year. These goals also provide pointers
for causal analysis, thus providing opportunities for Product and Process improvement.
The data is collected from Project Trackers maintained by individual projects. The statistical
analysis of all this data has been presented in the report along with inferences.
Some of the data points in almost all metrics in the report fall outside the control limits of process
performance, as indicated by the statistical analysis using Box Plots. They have not been
considered fit for the analysis but are included in percentile calculation.
Please refer to Appendix-A for the list of defaulting projects.
Please refer to Appendix-B for Rules for defining Control Limits.
2.1 Objectives
The purpose of this report is:
To provide basis for estimation planning.
To develop and sustain a measurement culture in the organization to support information
needs of the management.
To determine the process capability of the organizations’ standard process based on the
process performance data collected from projects.
To provide a basis to the projects to establish and improve their process performance
goals and to analyze the performance of project level process as defined in respective PMPs
and/or PDPs.
To provide basis to the projects to establish project performance goals based on
organization’s strategies and plans, and adjust project level process to accomplish the
defined goals.
To enable the Measurement Council & SEPG to take decisions on setting organizational
goals.
2.2 Scope
All projects being implemented are covered by this report unless a waiver have been obtained in
project PDP. This report analyses data for the period Apr–June, 2009.
LCL The minimum value until which the process can go in a controlled
Page 10 of 15
XYZ Company QMG R-OPP-PPBR-Q22009
Process Performance Baseline Report
April to June, 2009
v1.2
manner. Any value within the lower and the upper control limit means
the process is controlled and stable.
Mean The mean value is simply the average of all data points applicable.
PA Process Area
P50 / Median 50th percentile i.e. the middle value of the data set.
UCL The maximum value until which the process can go in a controlled and
stable.
UB and LB The Upper and the Lower Bounds. These are the limits specified by the
client, management or the statute.
Note: For other terms and acronyms, please refer to CMMI v 1.2 glossary and acronyms.
Page 11 of 15
XYZ Company QMG R-OPP-PPBR-Q22009
Process Performance Baseline Report
April to June, 2009
v1.2
4 ORGANIZATIONAL GOALS
Based on the analysis of the data received from projects, following statistical analysis has been
done by the measurement Council:
4.1 Maintenance Projects:
Eighty Maintenance projects are considered for this analysis.
Metrics Goal (Oct - Dec’ 07) - Maintenance Projects
Metrics Benchmark Performanc Goal * UCL* LCL*
e Mean
% Schedule Variance 9 2.72 5, +15, -20 +20 -15
(Original)
% Schedule Variance 9 0.56 1, +10, -10 +11 -9
(Revised)
% Schedule Slippage 9 1.99 1, +10, -10 +11 -9
(Revised)
% Effort Variance 26 0.68 5, + 15, - 15 +20 -10
% Effective Project NA 85.95 90, +10, -5 +100 +85
Utilization
% QA Effectiveness 96 61.12 80, +15, -10 +95 +70
Defect Rate (Defects / NA 1 50, +30, -30 +80 +20
Person Month)
Delivered Defect Rate NA 0.47 4, +2, -2 +6 +2
(defects / Person month)
% Rework Effort NA 3.47 10, +5, -5 +15 +5
* Already published goals and control limits are reproduced here.
Page 12 of 15
XYZ Company QMG R-OPP-PPBR-Q22009
Process Performance Baseline Report
April to June, 2009
v1.2
5 MAINTENANCE PROJECTS
Refer the attached document for details.
MAINTENANCE
PROJECTS
Page 13 of 15
XYZ Company QMG R-OPP-PPBR-Q22009
Process Performance Baseline Report
April to June, 2009
v1.2
ENHANCEMENT AND
DEVP PROJECTS
7 SUPPORT PROJECTS
Refer the attached document for details.
<TBD>
8 QA (TESTING) PROJECTS
Refer the attached document for details.
<TBD>
9 RECOMMENDATIONS
1. The timeliness and the correctness of the data in project trackers need to drastically improve.
Making the people who are responsible for projects revenue (AMCs or PMs), accountable for
the quality of data in project trackers seems to be the only solution of this problem.
2. Measurement Task Force will evolve suitable Process Performance Models, e.g. complexity
model, system dynamics modes or defect prediction models e.t.c. after 2 or 3 PPBR
releases, when the data is more reliable.
3. Include PPBRs and measurement council’s proceedings under the scope of internal audits.
4. The inferences in future PPBRs shall be based on further investigation of facts and soft
issues- not only on the statistics. This remains the responsibility of Measurement Council.
5. Once the quality of data improves, the council will start analyzing data grouped in various
buckets to enable apple to apple comparisons across the organization.
6. It was decided during the Fagan Inspection that effective project’s utilization is a project level
or contractual requirement and it doesn’t help to measure it at organizational level. This may
also be noted that manpower utilization is being tracked every month through time sheet
system.
Page 14 of 15
XYZ Company QMG R-OPP-PPBR-Q22009
Process Performance Baseline Report
April to June, 2009
v1.2
2. Levels of Exclusion:
a. Criteria 1: Null/Blank data points are excluded from analysis. Only exception to this rule
is DR for development, maintenance and enhancement projects.
b. Criteria 2: Outliers identified using Box Plot method are considered as those pertaining
to special causes of variations and are therefore excluded from deriving UCL/LCL.
Page 15 of 15