100% found this document useful (1 vote)
3K views27 pages

Sample Master Test Plan

This document is a master test plan for a company's product improvement project. It outlines the scope of testing, including the functions to be tested and not tested. The test approach involves functional testing throughout development by QA and business analysts, with formal testing at the end of iterations by subject matter experts. Special testing tools like Microsoft Visual Studio will be used and configured. Regression, performance, security and other types of testing are also defined.
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOC, PDF, TXT or read online on Scribd
100% found this document useful (1 vote)
3K views27 pages

Sample Master Test Plan

This document is a master test plan for a company's product improvement project. It outlines the scope of testing, including the functions to be tested and not tested. The test approach involves functional testing throughout development by QA and business analysts, with formal testing at the end of iterations by subject matter experts. Special testing tools like Microsoft Visual Studio will be used and configured. Regression, performance, security and other types of testing are also defined.
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOC, PDF, TXT or read online on Scribd
You are on page 1/ 27

MASTER TEST PLAN

Version: 1.0

Prepared for

[Enter Company Name Here]

[Name] Improvement Project


Revision History

Date Version Description Author


TABLE OF CONTENTS

1 INTRODUCTION...............................................................................................5
1.1 PURPOSE............................................................................................................................5
1.2 SCOPE...............................................................................................................................6
1.2.1 Product Description...............................................................................................7
1.2.2 The Current Product Version..................................................................................7
1.2.3 Updating the Product.............................................................................................8
1.2.4 High-Level Product Development Objective...........................................................8
1.3 TEST PLAN OBJECTIVES...........................................................................................................9
2 TEST ITEMS (FUNCTIONS)..............................................................................10
2.1 CLIENT APPLICATION.............................................................................................................10
2.2 QUICK HELP TESTING...........................................................................................................10
2.3 LICENSE KEY......................................................................................................................11
2.4 SECURITY..........................................................................................................................11
3 SOFTWARE RISK ISSUES................................................................................11
3.1 SCHEDULE.........................................................................................................................11
3.2 TECHNICAL........................................................................................................................11
3.3 MANAGEMENT.....................................................................................................................11
3.4 PERSONNEL.......................................................................................................................12
3.5 REQUIREMENTS...................................................................................................................12
4 FUNCTIONS TO BE TESTED.............................................................................12
5 FUNCTIONS TO NOT BE TESTED......................................................................12
6 TEST APPROACH............................................................................................13
FUNCTIONAL TESTING WILL BE CONDUCTED DURING THE ENTIRE APPLICATION DEVELOPMENT LIFE CYCLE BY THE QUALITY
ASSURANCE ENGINEER AND THE SYSTEM BUSINESS ANALYST. AT THE CONCLUSION OF EACH ITERATION, FORMAL TESTING
WILL BE CONDUCTED BY THE BUSINESS UNIT SUBJECT MATTER EXPERTS IN TWO CYCLES; WHILE OVERALL TESTING OF THE

ENTIRE SYSTEM WILL BE CONDUCTED BY THE QUALITY ASSURANCE ENGINEER AND THE SYSTEM BUSINESS ANALYST IN ONE

FINAL CYCLE.............................................................................................................................13

6.1 SPECIAL TESTING TOOLS........................................................................................................13


6.2 SPECIAL TRAINING ON TESTING TOOLS........................................................................................14
SINCE MICROSOFT VISUAL STUDIO 2010 ULTIMATE IS A RELATIVELY NEW TOOL ON THE MARKET, SOME SPECIAL TRAINING
WILL BE REQUIRED. DUE TO THE CURRENT TIME CONSTRAINTS OF THE PROJECT, THIS TRAINING WILL BE SCHEDULED AND

COMPLETED AT A LATER DATE.........................................................................................................14

HOWEVER, THE DEVELOPMENT TEAM; INCLUDING THE SQA ENGINEER, HAS PROACTIVELY BEGUN “SELF-LEARNING” AND
UTILIZING THE TOOL AND HAS INTEGRATED THE USE OF THE MICROSOFT TOOLS INTO ITS OVERALL TESTING STRATEGY.......14

6.3 TEST METRICS....................................................................................................................14


TESTING INFORMATION WILL BE COLLECTED AND ORIENTED TOWARD THE LEVEL OF TESTING. FOR HIGHER LEVELS,
APPLICATION AND FUNCTIONAL DATA WILL BE COLLECTED AND DOCUMENTED. FOR LOWER LEVELS, PROGRAM, UNIT, MODULE

AND BUILD DATA WILL BE COLLECTED AND DOCUMENTED...........................................................................14

ALL TEST RESULTS WILL BE DOCUMENTED IN THE MASTER TEST CASE SPREADSHEET, REVIEWED BY THE BUSINESS UNIT TEAM
LEADER, AND FORWARDED TO THE QUALITY ASSURANCE ENGINEER AND THE SYSTEM BUSINESS ANALYST...................14
THE TEST RESULTS WILL BE EVALUATED AND ENTERED INTO THE TEAM FOUNDATION SERVER THROUGH THE USE OF VISUAL
STUDIO TEST MANAGER; WHERE EACH “AUTOMATABLE” TEST CASE WILL BE RECORDED AND RUN USING THE “CODED UI
TEST” FEATURE IN VISUAL STUDIO 2010 ULTIMATE. ............................................................................14
REPORTS WILL BE GENERATED HIGHLIGHTING:......................................................................................14
THE NAME OF THE ASSIGNED TESTER................................................................................................14
THE NUMBER OF TEST CASES COMPLETED...........................................................................................14
THE NUMBER OF PASSED AND FAILED TEST CASES..................................................................................14
THE NUMBER OF OPEN ISSUED (BY PRIORITY).......................................................................................15
EVENTUALLY, PERFORMANCE TESTING WILL BE CONDUCTED TO MEASURE THE SYSTEM RESPONSE TIME, BASELINE LOAD, AND
SYSTEM STRESS CAPACITY..............................................................................................................15

6.4 CONFIGURATION MANAGEMENT.................................................................................................15


TEST CONFIGURATION MANAGEMENT WILL BE MANAGED BY THE LAB MANAGER TOOL; INCLUDED WITH THE MICROSOFT
VISUAL STUDIO 2010 ULTIMATE PACKAGE.........................................................................................15
LAB MANAGER CAN FULLY PROVISION AND READY MULTIPLE ENVIRONMENTS FOR TESTING SO THAT BUILD SCRIPTS CAN
EXPLICITLY TARGET A PARTICULAR LAB CONFIGURATION AT BUILD TIME. LAB MANAGEMENT STORES THE ENVIRONMENTS AS

VIRTUAL MACHINE IMAGES IN A LIBRARY OF PRE-BUILT IMAGES USING SYSTEM CENTER VIRTUAL MACHINE MANAGER

(SCVMM) TO ENSURE TEAMS ALWAYS BEGIN THEIR TESTING FROM A KNOWN CONFIGURATION. ..............................15
THE FOLLOWING OPERATING SYSTEMS AND BROWSER WILL BE USED IN MULTIPLE COMBINATIONS FOR TESTING THE [PROJECT
NAME] APPLICATION:...................................................................................................................15
OPERATING SYSTEMS.................................................................................................................15
WINDOWS XP.........................................................................................................................15
WINDOWS VISTA......................................................................................................................15
WINDOWS 7...........................................................................................................................15
BROWSERS.............................................................................................................................15
FIREFOX 3.0..........................................................................................................................15
INTERNET EXPLORER 7.0............................................................................................................15
INTERNET EXPLORER 8.0............................................................................................................15
6.5 REGRESSION TESTING............................................................................................................15
REGRESSION TESTING WILL BE CONDUCTED AT THE CONCLUSION OF EACH TESTING ITERATION BY THE SQA DEPARTMENT
AND ANY ASSIGNED BUSINESS UNIT SUBJECT MATTER EXPERTS. IN MOST CASES, THE TESTING WILL BE BASED ON SEVERITY OF

DEFECTS DETECTED.....................................................................................................................16

6.6 REQUIREMENTS MANAGEMENT..................................................................................................16


THE BUSINESS REQUIREMENTS WILL BE ELICITED AND MANAGED BY THE SYSTEMS BUSINESS ANALYST. ANY ELEMENTS IN
THE REQUIREMENTS AND DESIGN THAT DO NOT MAKE SENSE OR ARE NOT TESTABLE WILL BE IMMEDIATELY DOCUMENTED AND

REPORTED TO THE SBA; WHO WILL IN TURN ADDRESS THESE ISSUES WITH THE SHAREHOLDERS FOR FURTHER CLARIFICATION.
..........................................................................................................................................16
7 TEST STRATEGY.............................................................................................16
7.1 SYSTEM TEST.....................................................................................................................16
7.2 PERFORMANCE TEST.............................................................................................................16
7.3 SECURITY TEST...................................................................................................................16
7.4 AUTOMATED TEST................................................................................................................16
7.5 STRESS AND VOLUME TEST ...................................................................................................16
7.6 RECOVERY TEST ..............................................................................................................17
7.7 DOCUMENTATION TEST .......................................................................................................17
7.8 BETA TEST.......................................................................................................................17
7.9 USER ACCEPTANCE TEST........................................................................................................17
8 ENTRY AND EXIT CRITERIA.............................................................................17
8.1 TEST PLAN........................................................................................................................17
8.1.1 Test Plan Entry Criteria........................................................................................17
8.1.2 Test Plan Exit Criteria..........................................................................................17
8.1.3 Suspension and Resumption Criteria...................................................................17
8.2 TEST CYCLES.....................................................................................................................18
8.2.1 Test Cycle Entry Criteria......................................................................................18
8.2.2 Test Cycle Exit Criteria........................................................................................18
9 DELIVERABLES..............................................................................................18
9.1 TEST EVALUATION SUMMARIES..................................................................................................19
9.2 INCIDENT LOGS AND CHANGE REQUESTS......................................................................................19
10 ENVIRONMENT REQUIREMENTS....................................................................19
10.1 BASE SYSTEM HARDWARE....................................................................................................19
10.2 BASE SOFTWARE ELEMENTS IN THE TEST ENVIRONMENT..................................................................20
10.3 PRODUCTIVITY AND SUPPORT TOOLS.........................................................................................20
11 RESPONSIBILITIES, STAFFING AND TRAINING NEEDS.....................................20
11.1 PEOPLE AND ROLES............................................................................................................20
11.2 STAFFING AND TRAINING NEEDS..............................................................................................23
12 TEST SCHEDULE...........................................................................................23
13 POTENTIAL RISKS AND CONTINGENCIES........................................................24
14 CONTROL PROCEDURES...............................................................................24
14.1 REVIEWS........................................................................................................................24
14.2 BUG REVIEW MEETINGS........................................................................................................24
14.3 CHANGE REQUEST.............................................................................................................24
14.4 DEFECT REPORTING............................................................................................................25
15 DOCUMENTATION........................................................................................25
16 ITERATION MILESTONES...............................................................................26
17 MANAGEMENT PROCESS AND PROCEDURES..................................................27
17.1 PROBLEM REPORTING, ESCALATION, AND ISSUE RESOLUTION.............................................................27
17.2 APPROVAL AND SIGNOFF......................................................................................................27

1 Introduction

1.1 Purpose
The purpose of this Software Quality Assurance (SQA) Plan is to establish the goals,
processes, and responsibilities required to implement effective quality assurance
functions for the [Project Name] Improvement project.

This [Project Name] Improvement Software Quality Assurance Plan provides the
framework necessary to ensure a consistent approach to software quality assurance
throughout the project life cycle. It defines the approach that will be used by the
Software Quality (SQ) personnel to monitor and assess software development processes
and products to provide objective insight into the maturity and quality of the software.
The systematic monitoring of the [Project Name] products, processes, and services will
be evaluated to ensure they meet requirements and comply with [COMPANY NAME] and
[Project Name] policies, standards, and procedures, as well as applicable Institute of
Electrical and Electronic Engineers (IEEE) standards.

The overall purpose of this Master Test Plan is to gather all of the information necessary
to plan and control the test effort for testing the [Project Name] application. It
describes the approach to testing the software, and will be the top-level plan used by
testers to direct the test effort.

This plan is designed to create clear and precise documentation of the test methods and
processes that [COMPANY NAME] will use throughout the course of the [Project Name]
system verification testing.

This plan covers SQA activities throughout the formulation and implementation phases
of the [Project Name] mission. SQA activities will continue through operations and
maintenance of the system.

This Documenting of the test methods and processes will serve as the basis for ensuring
that all major milestones and activities required for effective verification testing can
efficiently and successfully be accomplished. This Master Test Plan will be modified and
enhanced as required throughout the verification testing engagement.

1.2 Scope
The scope of this quality assurance effort is to validate the full range of activities related
to the functionality of the flagship product of the [COMPANY NAME]: the [Project Name]
Course; as it undergoes re-designing and re-building.

This test plan describes the unit, subsystem integration, and system level tests that will
be performed on components of the [Name] application. It is assumed that prior to
testing each subsystem to be tested will have undergone an informal peer review and
only code that has successfully passed a peer review will be tested.

Unit tests will be done through test driver program that perform boundary checking and
basic black box testing.

The scope of this test effort outlines the quality assurance methodology, process and
procedures used to validate system functionality and the user’s ability to navigate
through the 4 major components of the [Project Name] Course solution:
 The Prep Course
 Practice Exams
 The [PRODUCT NAME]Exam
 A Progress Tracking

The functional testing will include a Prep Course creation and maintenance tool (Admin
Tool) for use by internal [COMPANY NAME] staff to facilitate creating new Prep Course,
Practice Exam, and [PRODUCT NAME]Exam question and answer databases, as well as
testing the ability to maintain those question and answer databases by adding new
questions, correcting errors in existing questions, modifying responses, etc. This
internal tool will support the maintenance of multiple versions of the software, and
should be capable of generating a database suitable for distribution with the software
itself. Finally, the test effort will include an internal tool which allows [COMPANY NAME]
staff to review results that are uploaded by users for the Prep Course, Practice Exams,
and [PRODUCT NAME]Exam.

1.2.1 Product Description


The [Project Name] Course is a software product sold by the [COMPANY NAME] to
its members to help them study and prepare for the [PRODUCT NAME]Exam, which
is required for an individual to earn their Certified Fraud Examiner certification. The
[Project Name] Course software actually encompasses two different applications:
The [Project Name] Course itself, which offers over a thousand sample questions for
the member to study, and the actual [PRODUCT NAME]Exam, which administers
randomly-selected questions from a question database in a timed session.

The [Project Name] Course and [PRODUCT NAME]Exam are offered in US,
Canadian, UK, and International editions, and a new revision of each is released
annually.

1.2.2 The Current Product Version


Currently, this product exists in multiple buyable formats, including the stand-alone
Prep Course and Exam on CD, a downloadable version of the Prep + Exam, the
[PRODUCT NAME]Exam-only CD, a downloadable version of the [PRODUCT
NAME]Exam, and the [Project Name] Toolkit, which includes hard-copy study
material. In each case, both the [Project Name] Course and the [PRODUCT
NAME]Exam software programs are Microsoft Windows-based desktop applications
that utilize an encrypted Microsoft Access database for data storage. Users of the
software are initially provided with a license key to unlock the Prep Course.

Upon successful completion of the Prep Course, they are then provided a second
key to unlock the [PRODUCT NAME]Exam. The user may submit the results of the
Exam by exporting an encrypted text file from the software and then emailing that
file to the [COMPANY NAME] for grading.
1.2.3 Updating the Product
The software has proven to be a very stable and reliable product despite the fact
that it was developed in 1999; however, the product is showing its age and needs
to be re-written to bring it up to current standards. Since the product has proven to
be very successful and has received a great deal of positive feedback from users, it
is the intent of this project to keep most of the features of the existing version
intact. Additionally, several new features and enhancements will be added to
improve the customer experience with the product, help users work through the
Prep Course more efficiently, as well as to enhance the installation, maintenance,
and administration of the software. For this initial release, the product will continue
to be offered as an installable product on Windows-based PCs and will not require
Internet access, with the exception of a few specific tasks (such as license key
verification, transmitting exam results to the [COMPANY NAME], and receiving
updates and patches).

1.2.4 High-Level Product Development Objective


The high-level objectives of this project are as follows:

 Update the software to a current technology platform while


maintaining the core functionality of the original product

 Re-design the user interface to take advantage of current technology


standards, and to deliver a better user experience

 Improve the process for creating new versions of the [Name] by


providing tools that allow the addition, modification, and removal of questions
from the questions database, maintenance of sections and sub-sections, and
error-checking routines that ensure the question and answers database is
assembled correctly

 Provide a method for delivering updates to the product electronically


and seamlessly to the user

 Provide a feature set that helps the user proceed through the
certification process faster and more efficiently, from the time the product is
purchased to the time the exam results are submitted to the [COMPANY
NAME] for evaluation

 Allow internal non-technical personnel to manage the creation and


maintenance of new [Name] versions
1.3 Test Plan Objectives
This Test Plan for the new [Project Name] Course solutions supports the following
objectives:

This Test Plan supports the following objectives:

 Outlines and defines the overall test approach that will be used.

 Identifies hardware, software, and tools to be used to support the testing


efforts.

 Defines the types of tests to be performed.

 Defines the types of exam data required for effective testing.

 Defines the types of security threats and vulnerabilities against which each
exam system will be tested.

 Identifies and establishes traceability from the Requirements Matrix to test


cases and from test cases to the Requirements Matrix.

 Serves as a foundation for the development of Test Plans and Test Cases.

 Defines the process for recording and reporting test results.

 Defines the process for regression testing and closure of discrepancies.

 Identifies the items that should be targeted by the tests.

 Identifies the motivation for and ideas behind the test areas to be covered.

 Identifies the required resources and provides an estimate of the test efforts.

 List the deliverable elements of the test activities.

 Define the activities required to prepare for and conduct System, Beta and
User Acceptance testing

 Communicate to all responsible parties the System Test strategy

 Define deliverables and responsible parties

 Communicate to all responsible parties the various Dependencies and Risks

 Scope
2 Test Items (Functions)
The testing effort will be concentrated on the following functions of the application:

2.1 Client Application


 Installation
 Uninstall
 Activation
 Prep Course Home Page
 Pre-Assessment
 Review Sessions
 Practice Exams
 Exam Application
 Admin Tool
 Editions
 Revisions
 Sections
 Subsections
 Topics
 Demo Questions (create, view, edit, and approve)
 Test Search Capability
 License Keys
 Exam Keys
 Review Exam Key Requests
 Review Exam Submissions
 Issue Reports

2.2 Quick Help Testing


The system will provide the [Name] users the ability to access a help menu; consisting
of links to various options. These following utility options will be tested:

 [COMPANY NAME] Reference Manual


 Launch [Project Name] Course
 Your Contact Information
 [COMPANY NAME].com
 FAQs
 Online Discussion Forums
 [PRODUCT NAME]Certification
 Qualification Checklist
 Request Exam Key
 Enter Exam Key
 Submit [PRODUCT NAME]Exam to [COMPANY NAME]
 Utilities
 Check for Updates
 Activate Product
 Report Issue

2.3 License Key


Current the License Key is requested by submitting an email to the [COMPANY NAME]
Certification Team. The [COMPANY NAME] Certification Team will electronically generate
and email the license key back to the requestor.

New functionality will be implemented to enhance this process. The users will be
required to complete and electronically submit a License Key request form to the
[COMPANY NAME] Certification Team. The [COMPANY NAME] Certification Team will
electronically generate and email the license key back to the requestor.

This new functionality will be included in the scope of the testing effort.

2.4 Security
Each user will need a UserId and password to login to the system. The UserId and
password will be created by the user during the registration process. The system should
validate that the UserId and password both meet the correct format standard. Once the
registration form has been electronically submitted, the system will notify the user that
the requested identification information has been accepted and the request has been
granted. The system will require the users to change the password every 30 days.

3 Software Risk Issues

3.1 Schedule
The schedule for each phase is very aggressive and could affect testing. A slip in the
schedule in one of the other phases could result in a subsequent slip in the test phase.
Close project management is crucial to meeting the forecasted completion date.

3.2 Technical
Since this is a new [Name] system, in the event of a failure the old system can be used.
We will run our test in parallel with the production system so that there is no downtime
of the current system.

3.3 Management
Management support is required so when the project falls behind, the test schedule does
not get squeezed to make up for the delay. Management can reduce the risk of delays
by supporting the test team throughout the testing phase and assigning people to this
project with the required skills set.

3.4 Personnel
Due to the aggressive schedule, it is very important to have experienced testers on this
project. Unexpected turnovers can impact the schedule. If attrition does happen, all
efforts must be made to replace the experienced individual.

3.5 Requirements
The test plan and test schedule are based on the current Requirements Document. Any
changes to the requirements could affect the test schedule and will need to be approved
by the CCB.

4 Functions to Be Tested

The following is a list of functions that will be tested:

 Add/update user information


 Search / Lookup employee information
 Escape to return to Main Menu
 Security features
 Error messages
 Prep Course Functionality
 New License Key Request Functionality
 Practice Exam Functionality
 [PRODUCT NAME]Exam Functionality
 Progress Tracking Functionality
 Screen mappings (GUI flow). Includes default settings

A Requirements Validation Matrix will “map” the test cases back to the requirements. See
Deliverables.

5 Functions to Not Be Tested

This is to be determined.
6 Test Approach
Functional testing will be conducted during the entire Application Development Life Cycle by
the Quality Assurance Engineer and the System Business Analyst. At the conclusion of each
iteration, formal testing will be conducted by the business unit subject matter experts in two
cycles; while overall testing of the entire system will be conducted by the Quality Assurance
Engineer and the System Business Analyst in one final cycle.
The overall testing approach of the project will address and encompass the following rules
and processes:

6.1 Special Testing Tools

 Microsoft Visual Studio 2010 Ultimate - The comprehensive suite of application


lifecycle management tools used by the [COMPANY NAME] Software Development
Team to ensure quality results, from design to deployment. It includes: Integrated
Development Environment, Development Platform Support, Team Foundation Server,
MSDN Subscription, Testing Tools, Database Development, Debugging and
Diagnostics, Application Lifecycle Management, Architecture and Modeling, and Lab
Management.

 Visual Studio Test Professional 2010 - An integrated testing toolset that delivers
a complete plan-test-track workflow for in-context collaboration between testers and
developers.

 Microsoft Visual Studio Team Foundation Server 2010 (TFS) - The


collaboration platform at the core of our application lifecycle management (ALM)
process. Team Foundation Server 2010 automates the software delivery process and
enables everyone on our team to collaborate more effectively, be more agile, and
deliver better quality software while building and sharing institutional knowledge.
Project artifacts like requirements, tasks, bugs, source code, and build and test
results are stored in a data warehouse. The tool also contains the reporting,
historical trending, full traceability, and real-time visibility into quality and progress.

6.2 Special Training on Testing Tools

Since Microsoft Visual Studio 2010 Ultimate is a relatively new tool on the market, some
special training will be required. Due to the current time constraints of the project, this
training will be scheduled and completed at a later date.

However, the Development Team; including the SQA Engineer, has proactively begun
“self-learning” and utilizing the tool and has integrated the use of the Microsoft tools
into its overall testing strategy.

6.3 Test Metrics

Testing Information will be collected and oriented toward the level of testing. For higher
levels, application and functional data will be collected and documented. For lower
levels, program, unit, module and build data will be collected and documented.

All test results will be documented in the Master Test Case spreadsheet, reviewed by
the business unit team leader, and forwarded to the Quality Assurance Engineer and the
System Business Analyst.

The test results will be evaluated and entered into the Team Foundation Server through
the use of Visual Studio Test Manager; where each “automatable” test case will be
recorded and run using the “Coded UI Test” feature in Visual Studio 2010 Ultimate.

Reports will be generated highlighting:

 The name of the assigned tester

 The number of test cases completed

 The number of passed and failed test cases


 The number of open issued (by priority)

Eventually, performance testing will be conducted to measure the system response


time, baseline load, and system stress capacity.

6.4 Configuration Management

Test configuration management will be managed by the Lab Manager tool; included with
the Microsoft Visual Studio 2010 Ultimate package.

Lab Manager can fully provision and ready multiple environments for testing so that
build scripts can explicitly target a particular lab configuration at build time. Lab
Management stores the environments as virtual machine images in a library of pre-built
images using System Center Virtual Machine Manager (SCVMM) to ensure teams always
begin their testing from a known configuration.

The following operating systems and browser will be used in multiple combinations for
testing the [Project Name] application:

Operating Systems

 Windows XP

 Windows Vista

 Windows 7

Browsers

 Firefox 3.0

 Internet Explorer 7.0

 Internet Explorer 8.0

6.5 Regression Testing


Regression testing will be conducted at the conclusion of each testing iteration by the
SQA Department and any assigned business unit subject matter experts. In most cases,
the testing will be based on severity of defects detected.

6.6 Requirements Management

The business requirements will be elicited and managed by the Systems Business
Analyst. Any elements in the requirements and design that do not make sense or are
not testable will be immediately documented and reported to the SBA; who will in turn
address these issues with the shareholders for further clarification.

7 Test Strategy

The test strategy consists of a series of different tests that will fully exercise the [Name]
system. The primary purpose of these tests is to uncover the systems limitations and
measure its full capabilities. A list of the various planned tests and a brief explanation
follows below.

7.1 System Test


The System tests will focus on the behavior of the [Name] system. User scenarios will be
executed against the system as well as screen mapping and error message testing.
Overall, the system tests will test the integrated system and verify that it meets the
requirements defined in the requirements document.

7.2 Performance Test


Performance test will be conducted to ensure that the [Name] system’s response times
meet the user expectations and do not exceed the specified performance criteria. During
these tests, response times will be measured under heavy stress and/or volume.

7.3 Security Test


Security tests will determine how secure the new [Name] system is. The tests will verify
that unauthorized user access to confidential data is prevented.

7.4 Automated Test


A suite of automated tests will be developed to test the basic functionality of the [Name]
system and perform regression testing on areas of the systems that previously had
critical/major defects. The tool will also assist us by executing user scenarios thereby
emulating several users.

7.5 Stress and Volume Test


We will subject the [Name] system to high input conditions and a high volume of data
during the peak times. The System will be stress tested using twice (20 users) the
number of expected users.

7.6 Recovery Test


Recovery tests will force the system to fail in a various ways and verify the recovery is
properly performed. It is vitally important that all [Name] data is recovered after a
system failure & no corruption of the data occurred.

7.7 Documentation Test


Tests will be conducted to check the accuracy of the user documentation. These tests
will ensure that no features are missing, and the contents can be easily understood.

7.8 Beta Test


The [Name] department will beta tests the new [Name] system and will report any
defects they find. This will subject the system to tests that could not be performed in our
test environment.

7.9 User Acceptance Test


Once the [Name] system is ready for implementation, the [Name] department will
perform User Acceptance Testing. The purpose of these tests is to confirm that the
system is developed according to the specified user requirements and is ready for
operational use.

8 Entry and Exit Criteria

8.1 Test Plan


8.1.1 Test Plan Entry Criteria
Package coding is complete and the code has been informally reviewed by the team.

8.1.2 Test Plan Exit Criteria


Either the second semester ends or all of the use case requirements have been
verified.

8.1.3 Suspension and Resumption Criteria


Testing will be suspended on critical design flaws that will require a package,
package interface redesigned. Testing will resume when the coding is complete and
code is reviewed successfully.
If testing is suspended, resumption will only occur when the problem(s) that caused
the suspension has been resolved. When a critical defect is the cause of the
suspension, the “FIX” must be verified by the test department before testing is
resumed.

8.2 Test Cycles


8.2.1 Test Cycle Entry Criteria
TBD

8.2.2 Test Cycle Exit Criteria


All tests specified at the start of the iteration have completed successfully, or the end
of the second semester occurs.

If any defects are found which seriously impact the test progress, the QA manager
may choose to suspend testing. Criteria that will justify test suspension are:
Hardware/software is not available at the times indicated in the project schedule.

 Source code contains one or more critical defects, which seriously prevents or
limits testing progress.

 Assigned test resources are not available when needed by the test team.

9 Deliverables

Deliverable Responsibility Delivery Date

Develop Test cases Testers 10/13/2010


Test Case Review Test Lead, Dev. Lead, 10/15/2010
Testers
Develop Automated test Testers TBD
suites
Requirements Validation Test Lead TBD
Matrix
Execute manual and Testers & Test Lead 9/20 – 10/15/2010
automated tests
Complete Defect Reports Everyone testing the product On-going
Document and Test Lead Weekly
communicate test
status/coverage
Execute Beta tests [Name] Department Users 10/18 – 10/29/2010
Document and [Name] Department Manager TBD
communicate Beta test
status/coverage
Execute User Acceptance [Name] Department Users 10/18 – 10/29/2010
tests
Document and [Name] Department Manager TBD
communicate Acceptance
test status/coverage
Final Test Summary Report Test Lead 11/1/2010

The following artifacts will be testing deliverables, available to the stakeholders:

9.1 Test Evaluation Summaries


These summaries will outline the tests conducted and their results.

9.2 Incident Logs and Change Requests


Incident log entries will be made for all bugs found during testing. The log will be used
to track the status of the bugs.
Any modifications to the requirements must be done through change requests, which
will ensure the proposed change is fully reviewed before being incorporated into the
[Name] application.

10 Environment Requirements

This section presents the non-human resources required for the Test Plan.

10.1 Base System Hardware


The following table sets forth the system resources for the test effort presented in this Test
Plan.

System Resources
Resource Quantity Name and Type
Application Server 1
System Resources
Resource Quantity Name and Type
—CPUs
—Memory
—Hard Disk 1
—Hard Disk 2
—Server Name
—IP Address
Test Development PCs TBD

10.2 Base Software Elements in the Test Environment


The following base software elements are required in the test environment for this Test
Plan.

Software Element Name Version Type and Other Notes


Windows Server
SQL Server

10.3 Productivity and Support Tools


The following tools will be employed to support the test process for this Test Plan.
Tool Category or Type Tool Brand Name Vendor or In-house Version
Test Management Microsoft 2010
Defect Tracking

11 Responsibilities, Staffing and Training Needs

This section outlines the personnel necessary to successfully test the [Name] application.
Since staff size is fixed these number may change.

11.1 People and Roles


This table shows the staffing assumptions for the test effort.
Human Resources

Role Minimum Resources Specific Responsibilities or


Recommended Comments

(number of full-time
roles allocated)

Test Analyst 1 Identifies and defines the specific tests


to be conducted.

Responsibilities include:

• identify test ideas

• define test details

• determine test results

• document change requests

• evaluate product quality

Test Designer 1 Defines the technical approach to the


implementation of the test effort.

Responsibilities include:

• define test approach

• define test automation architecture

• verify test techniques

• define testability elements

• structure test implementation


Human Resources

Role Minimum Resources Specific Responsibilities or


Recommended Comments

(number of full-time
roles allocated)

Tester 2 Implements and executes the tests.

Responsibilities include:

• implement tests and test suites

• execute test suites

• log results

• analyze and recover from test


failures

• document incidents

Database 1 Ensures test data (database)


Administrator, environment and assets are managed
Database Manager and maintained.

Responsibilities include:

• Support the administration of test


data and test beds (database).

Designer 1 Identifies and defines the operations,


attributes, and associations of the test
classes.

Responsibilities include:

• defines the test classes required


to support testability
requirements as defined by the
test team
Human Resources

Role Minimum Resources Specific Responsibilities or


Recommended Comments

(number of full-time
roles allocated)

Implementer 2 Implements and unit tests the test


classes and test packages.

Responsibilities include:

• creates the test components


required to support testability
requirements as defined by the
designer

11.2 Staffing and Training Needs


This section outlines how to approach staffing and training the test roles for the
project.

Staffing is fixed for the duration of this project. It is likely most of the staff will
assume some testing role.

12 Test Schedule

 Ramp up / System familiarization 10/01/10 - 10/15/10


 System Test 10/16/10 - 12/26/10
 Beta Test 12/28/10 - 01/18/11
 User Acceptance Test 01/19/11 - 02/01/11
13 Potential Risks and Contingencies

The overall risks to the project with an emphasis on the testing process are:

• Lack of personnel resources when testing is to begin.

• Lack of availability of required hardware, software, data or tools.

• Late delivery of the software, hardware or tools.

• Delays in training on the application and/or tools.

• Changes to the original requirements or designs.

Requirements definition will be complete by September 1, 2010, and, if the requirements


change after that date, the following actions will be taken:

• The test schedule and development schedule could move out an appropriate number
of days.

• The number of test performed could be reduced.

• The number of acceptable defects could be increased.

• Resources could be added to the test team.

• The test team could be asked to work overtime.

• The scope of the plan may be changed.

• There may be some optimization of resources.

14 Control Procedures

14.1 Reviews
The project team will perform reviews for each Phase. (i.e. Requirements Review,
Design Review, Code Review, Test Plan Review, Test Case Review and Final Test
Summary Review). A meeting notice, with related documents, will be emailed to each
participant.

14.2 Bug Review meetings


Regular weekly meeting will be held to discuss reported defects. The development
department will provide status/updates on all defects reported and the test department
will provide addition defect information if needed. All member of the project team will
participate.
14.3 Change Request
Once testing begins, changes to the [Name] system are discouraged. If functional
changes are required, these proposed changes will be discussed with the Change Control
Board (CCB). The CCB will determine the impact of the change and if/when it should be
implemented.

14.4 Defect Reporting


When defects are found, the testers will complete a defect report on the defect tracking
system and submit the report to the Eureka software engineers for analysis. The defect
tracking system is accessible by testers, developers & all members of the project team.
When a defect has been fixed or more information is needed, the Eureka developer will
change the status of the defect to indicate the current state. Once a defect is verified as
“fixed” by the testers; the testers will close the defect report.

15 Documentation
The following documentation will be available at the end of the test phase:

 Test Plan

 Test Cases

 Test Case review

 Requirements Validation Matrix

 Defect reports

 Final Test Summary Report


16 Iteration Milestones

Milestone Planned Actual Planned Actual


Start Date Start Date End Date End Date
Iteration Plan agreed
Iteration starts
Requirements base lined
Architecture base lined
User Interface base lined
First Build delivered to
test
First Build accepted into
test
First Build test cycle
finishes
[Build Two will not be
tested]
Third Build delivered to
test
Third Build accepted into
test
Third Build test cycle
finishes
Fourth Build delivered to
test
Fourth Build accepted
into test
Iteration Assessment
review
Iteration ends
17 Management Process and Procedures

17.1 Problem Reporting, Escalation, and Issue Resolution


Problems will be reported to the team by e-mail. Bugs will be listed on the project page
with the developer responsible for fixing each bug. Each bug will be given a priority,
which will determine when it is addressed in the current iteration.

17.2 Approval and Signoff


The team manager must approve the initial test plan. As tests are successfully
completed the testers will sign off use case requirements.

You might also like