Software Testing Q
Software Testing Q
(B) In Which Software Life Cycle Phase Does Testing Occur? OR (B) Can You Explain the PDCA Cycle and Where Testing Fits in?
Software testing is an important part of the software development process. In normal software development there are four important steps, also referred to, in short, as the PDCA (Plan, Do, Check, Act) cycle.
Plan: Define the goal and the plan for achieving that goal. Do/Execute: Depending on the plan strategy decided during the plan stage we do execution accordingly in this phase. Check: Check/Test to ensure that we are moving according to plan and are getting the desired results. Act: During the check cycle, if any issues are there, then we take appropriate action accordingly and revise our plan again.
So now to answer our question, where does testing fit in.you guessed it, the check part of the cycle. So developers and other stakeholders of the project do the "planning and building," while testers do the check part of the cycle.
(B) What is the Difference between White Box, Black Box, and Gray Box Testing?
Black box testing is a testing strategy based solely on requirements and specifications. Black box testing requires no knowledge of internal paths, structures, or implementation of the software being tested.
White box testing is a testing strategy based on internal paths, code structures, and implementation of the software being tested. White box testing generally requires detailed programming skills. There is one more type of testing called gray box testing. In this we look into the "box" being tested just long enough to understand how it has been implemented. Then we close up the box and use our knowledge to choose more effective black box tests. The following figure shows how both types of testers view an accounting application during testing. Black box testers view the basic accounting application. While during white box testing the tester knows the internal structure of the application. In most scenarios white box testing is done by developers as they know the internals of the application. In black box testing we check the overall functionality of the application while in white box testing we do code reviews, view the architecture, remove bad code practices, and do component level testing.
Wrong: The requirements have been implemented incorrectly. This defect is a variance from the given specification. Missing: There was a requirement given by the customer and it was not done. This is a variance from the specifications, an indication that a specification was not implemented, or a requirement of the customer was not noted properly. Extra: A requirement incorporated into the product that was not given by the end customer. This is always a variance from the specification, but may be an attribute desired by the user of the product. However, it is considered a defect because it's a variance from the existing requirements.
Note This question will be normally asked to see whether you can independently set up testing departments. Many companies still think testing is secondary. That's where a good testing manager should show the importance of testing. Bringing in the attitude of testing in companies which never had a formal testing department is a huge challenge because it's not about bringing in a new process but about changing the mentality. The following are the important steps used to define a testing policy in general. But it can change according to your organization. Let's discuss in detail the steps of implementing a testing policy in an organization. Definition: The first step any organization needs to do is define one unique definition for testing within the organization so that everyone is of the same mindset. How to achieve: How are we going to achieve our objective? Is there going to be a testing committee, will there be compulsory test plans which need to be executed, etc?. Evaluate: After testing is implemented in a project how do we evaluate it? Are we going to derive metrics of defects per phase, per programmer, etc. Finally, it's important to let everyone know how testing has added value to the project?. Standards: Finally, what are the standards we want to achieve by testing. For instance, we can say that more than 20 defects per KLOC will be considered below standard and code review should be done for it.
Figure 11: Establishing a testing policy The previous methodology is from a general point of view. Note that you should cover the steps in broader aspects.
(B) Should Testing Be Done Only After the Build and Execution Phases are Complete?
Note This question will normally be asked to judge whether you have a traditional or modern testing attitude. In traditional testing methodology (sad to say many companies still have that attitude) testing is always done after the build and execution phases. But that's a wrong way of thinking because the earlier we catch a defect, the more cost effective it is. For instance, fixing a defect in maintenance is ten times more costly than fixing it during execution.
Figure 12: Traditional way of testing Testing after code and build is a traditional approach and many companies have improved on this philosophy. Testing should occur in conjunction with each phase as shown in the following figure. In the requirement phase we can verify if the requirements are met according to the customer needs. During design we can check whether the design document covers all the requirements. In this stage we can also generate rough functional data. We can also review the design document from the architecture and the correctness perspectives. In the build and execution phase we can execute unit test cases and generate structural and functional data. And finally comes the testing phase done in the traditional way. i.e., run the system test cases and see if the system works according to the requirements. During installation we need to see if the system is compatible with the software. Finally, during the maintenance phase when any fixes are made we can retest the fixes and follow the regression testing.
(B) Are There More Defects in the Design Phase or in the Coding Phase?
Note This question is asked to see if you really know practically which phase is the most defect prone. The design phase is more error prone than the execution phase. One of the most frequent defects which occur during design is that the product does not cover the complete requirements of the customer. Second is wrong or bad architecture and technical decisions make the next phase, execution, more prone to defects. Because the design phase drives the execution phase it's the most critical phase to test. The testing of the design phase can be done by good review. On average, 60% of defects occur during design and 40% during the execution phase.
(B) What Kind of Input Do We Need from the End User to Begin Proper Testing?
The product has to be used by the user. He is the most important person as he has more interest than anyone else in the project. From the user we need the following data:
The first thing we need is the acceptance test plan from the end user. The acceptance test defines the entire test which the product has to pass so that it can go into production. We also need the requirement document from the customer. In normal scenarios the customer never writes a formal document until he is really sure of his requirements. But at some point the customer should sign saying yes this is what he wants. The customer should also define the risky sections of the project. For instance, in a normal accounting project if a voucher entry screen does not work that will stop the accounting functionality completely. But if reports are not derived the accounting department can use it for some time. The customer is the right person to say which section will affect him the most. With this feedback the testers can prepare a proper test plan for those areas and test it thoroughly. The customer should also provide proper data for testing. Feeding proper data during testing is very important. In many scenarios testers key in wrong data and expect results which are of no interest to the customer.
application first searches for the laser printer. If it finds a laser printer it uses the laser printer and prints it. If it does not find a laser printer, the application searches for dot matrix printer. If the application finds a dot matrix printer (DMP) the application prints using or an error is given. Now for whatever reason this application never searched for the dot matrix printer. So the application never got tested for the DMP. That means the exact conditions were never met for the DMP. This is called a latent defect. Now the same application has two defects: one defect is in the DMP search and the other defect is in the DMP print. But because the search of the DMP fails the print DMP defect is never detected. So the print DMP defect is a masked defect.
(B) A Defect Which Could Have Been Removed During the Initial Stage is Removed in a Later Stage. How Does this Affect Cost?
If a defect is known at the initial stage then it should be removed during that stage/phase itself rather than at some later stage. It's a recorded fact that if a defect is delayed for later phases it proves more costly. The following figure shows how a defect is costly as the phases move forward. A defect if identified and removed during the requirement and design phase is the most cost effective, while a defect removed during maintenance is 20 times costlier than during the requirement and design phases. For instance, if a defect is identified during requirement and design we only need to change the documentation, but if identified during the maintenance phase we not only need to fix the defect, but also change our test plans, do regression testing, and change all documentation. This is why a defect should be identified/removed in earlier phases and the testing department should be involved right from the requirement phase and not after the execution phase.
Figure 18: Workbench with phases and steps There are five tasks for every workbench:
Input: Every task needs some defined input and entrance criteria. So for every workbench we need defined inputs. Input forms the first steps of the workbench. Execute: This is the main task of the workbench which will transform the input into the expected output. Check: Check steps assure that the output after execution meets the desired result. Production output: If the check is right the production output forms the exit criteria of the workbench. Rework: During the check step if the output is not as desired then we need to again start from the execute step.
The following figure shows all the steps required for a workbench.
Figure 19: Phases in a workbench In real scenarios projects are not made of one workbench but of many connected workbenches. A workbench gives you a way to perform any kind of task with proper testing. You can visualize every software phase as a workbench with execute and check steps. The most important point to note is we visualize any task as a workbench by default we have the check part in the task. The following figure shows how every software phase can be visualized as a workbench. Let's discuss the workbench concept in detail:
Requirement phase workbench: The input is the customer's requirements; we execute the task of writing a requirement document, we check if the requirement document addresses all the customer needs, and the output is the requirement document. Design phase workbench: The input is the requirement document, we execute the task of preparing a technical document; review/check is done to see if the design document is technically correct and addresses all the requirements mentioned in the requirement document, and the output is the technical document. Execution phase workbench: This is the actual execution of the project. The input is the technical document; the execution is nothing but implementation/coding according to the technical document, and the output of this phase is the implementation/source code. Testing phase workbench: This is the testing phase of the project. The input is the source code which needs to be tested; the execution is executing the test case and the output is the test results. Deployment phase workbench: This is the deployment phase. There are two inputs for this phase: one is the source code which needs to be deployed and that is dependent on the test results. The output of this project is that the customer gets the product which he can now start using. Maintenance phase workbench: The input to this phase is the deployment results, execution is implementing change requests from the end customer, the check part is nothing but running regression testing after every change request implementation, and the output is a new release after every change request execution.
Alpha and beta testing has different meanings to different people. Alpha testing is the acceptance testing done at the development site. Some organizations have a different visualization of alpha testing. They consider alpha testing as testing which is conducted on early, unstable versions of software. On the contrary beta testing is acceptance testing conducted at the customer end. In short, the difference between beta testing and alpha testing is the location where the tests are done.
(I) Can You Explain the Concept of Defect Cascading? OR (B) Can You Explain How One Defect Leads to Other Defects?
Defect cascading is a defect which is caused by another defect. One defect triggers the other defect. For instance, in the accounting application shown here there is a defect which leads to negative taxation. So the negative taxation defect affects the ledger which in turn affects four other modules.
Usability testing is a testing methodology where the end customer is asked to use the software to see if the product is easy to use, to see the customer's perception and task time. The best way to finalize the customer point of view for usability is by using prototype or mock-up software during the initial stages. By giving the customer the prototype before the development start-up we confirm that we are not missing anything from the user point of view.
Pilot: The actual production system is installed at a single or limited number of users. Pilot basically means that the product is actually rolled out to limited users for real work. Gradual Implementation: In this implementation we ship the entire product to the limited users or all users at the customer end. Here, the developers get instant feedback from the recipients which allow them to make changes before the product is available. But the downside is that developers and testers maintain more than one version at one time. Phased Implementation: In this implementation the product is rolled out to all users in incrementally. That means each successive rollout has some added functionality. So as new functionality comes in, new installations occur and the customer tests them progressively. The benefit of this kind of rollout is that customers can start using the functionality and provide valuable feedback progressively. The only issue here is that with each rollout and added functionality the integration becomes more complicated. Parallel Implementation: In these types of rollouts the existing application is run side by side with the new application. If there are any issues with the new application we again move back to the old application. One of the biggest problems with parallel implementation is we need extra hardware, software, and resources.
The following figure shows the different launch strategies for a project rollout.
Figure 25: Requirement Traceability Note Many professionals still think testing is executing test cases on the application. But testing should be performed at all levels. In the requirement phase we can use the review and traceability matrix to check the validity of our project. In the design phase we can use the design review to check the correctness of the design and so on.
(B) How Do You Conclude Which Section is Most Risky in Your Application?
Note Here the interviewer is expecting a proper approach to rating risk to the application modules so that while testing you pay more attention to those risky modules, thus minimizing risk in projects. The following is a step by step approach for testing planning:
The first step is to collect features and concerns from the current documentation and data available from the requirement phase. For instance, here is a list of some features and concerns: Open table as spreadsheet
Table 1: Features and concerns Features Add a user Check user preferences Login user Add new invoice Print invoice Open table as spreadsheet Concerns Maintainability Security Performance
The table shows features and concerns. Features are functionalities which the end user will use, while concerns are global attributes of the project. For instance, the security has to be applied to all the features listed.
Once we have listed the features and concerns, we need to rate the probability/likelihood of failures in this feature. In the following section we have rated the features and concerns as low, high, and medium, but you can use numerical values if you want.
Once we have rated the failure probability, we need to rate the impact. Impact means if we make changes to this feature, how many other features will be affected? You can see in the following table that we have marked the impact section accordingly. We also need to define the master priority rating table depending on the impact and probability ratings. The following table defines the risk priority. Using the priority rating table we have defined priority for the following listed features. Depending on priority you can start testing those features first. Once the priority is set you can then review it with your team members to validate it. Open table as spreadsheet
Table 2: Probability rating according to features and concerns Features Add a user Check user preferences Login user Add new invoice Print invoice Open table as spreadsheet Concerns Maintainability Security Performance Probability of failure Low High High Low Low High Medium Probability of failure Low
Features Check user preferences Login user Add new invoice Print invoice Open table as spreadsheet Concerns Maintainability Security Performance
Probability of failure Low High High Table 4: Priority rating Open table as spreadsheet
Probability of failure Low Low Medium High Open table as spreadsheet Features Add a user Check user preferences Login user
Priority 1 1 2
Features Add new invoice Print invoice Open table as spreadsheet Concerns Maintainability Security Performance
Priority 4 3
Figure 27: Priority set according to the risk priority table The following figure shows the summary of the above steps. So list your concerns, rate the probabilities of failures, provide an impact rating, calculate risk/priority, and then review, review, and review.
project. For instance, one of the common exit criteria in projects is that the customer has successfully executed the acceptance test plan.
Requirement document: This document specifies what exactly is needed in the project from the customers perspective. Input from customer: This can be discussions, informal talks, emails, etc. Project plan: The project plan prepared by the project manager also serves as good input to finalize your acceptance test. In projects the acceptance test plan can be prepared by numerous inputs. It is not necessary that the above list be the only criteria. If you think you have something extra to add, go ahead.
Note
The following diagram shows the most common inputs used to prepare acceptance test plans.
(B) What's the Relationship between Environment Reality and Test Phases?
Environment reality becomes more important as test phases start moving ahead. For instance, during unit testing you need the environment to be partly real, but at the acceptance phase you should have a 100% real environment, or we can say it should be the actual real environment. The following graph shows how with every phase the environment reality should also increase and finally during acceptance it should be 100% real.
(B) What are Different Types of Verifications? OR (B) What's the Difference between Inspections and Walkthroughs?
As said in the previous sections the difference between validation and verification is that in validation we actually execute the application, while in verification we review without actually running the application. Verifications are basically of two main types: Walkthroughs and Inspections. A walkthrough is an informal form of verification. For instance, you can call your colleague and do an informal walkthrough to just check if the documentation and coding is correct. Inspection is a formal procedure and official. For instance, in your organization you can have an official body which approves design documents for any project. Every project in your organization needs to go through an inspection which reviews your design documents. If there are issues in the design documents, then your project will get a NC (nonconformance) list. You cannot proceed without clearing the NCs given by the inspection team.
(I) What is Coverage and What are the Different Types of Coverage Techniques?
Coverage is a measurement used in software testing to describe the degree to which the source code is tested. There are three basic types of coverage techniques as shown in the following figure:
Statement coverage: This coverage ensures that each line of source code has been executed and tested. Decision coverage: This coverage ensures that every decision (true/false) in the source code has been executed and tested. Path coverage: In this coverage we ensure that every possible route through a given part of code is executed and tested.
Note
While doing testing on the actual product, the code coverage testing tool is run simultaneously. While the testing is going on, the code coverage tool monitors the executed statements of the source code. When the final testing is completed we get a complete report of the pending statements and also get the coverage percentage.
Configuration management is the detailed recording and updating of information for hardware and software components. When we say components we not only mean source code. It can be tracking of changes for software documents such as requirement, design, test cases, etc. When changes are done in adhoc and in an uncontrolled manner chaotic situations can arise and more defects injected. So whenever changes are done it should be done in a controlled fashion and with proper versioning. At any moment of time we should be able to revert back to the old version. The main intention of configuration
management is to track our changes if we have issues with the current system. Configuration management is done using baselines. Note Please refer the baseline concept in the next question.
Figure 36: Baseline Baselines are very important from a testing perspective. Testing on a software product that is constantly changing will not get you anywhere. So when you actually start testing you need to first baseline the application so that what you test is for that baseline. If the developer fixes something then create a new baseline and perform testing on it. In this way any kind of conflict will be avoided.
There are a minimum of four test plan documents needed in any software project. But depending on the project and team members agreement some of the test plan documents can be deleted. Central/Project test plan: The central test plan is one of the most important communication channels for all project participants. This document can have essentials such as resource utilization, testing strategies, estimation, risk, priorities, and more. Acceptance test plan: The acceptance test plan is mostly based on user requirements and is used to verify whether the requirements are satisfied according to customer needs. Acceptance test cases are like a green light for the application and help to determine whether or not the application should go into production. System test plan: A system test plan is where all main testing happens. This testing, in addition to functionality testing, has also load, performance, and reliability tests. Integration testing: Integration testing ensures that the various components in the system interact properly and data is passed properly between them. Unit testing: Unit testing is done more on a developer level. In unit testing we check the individual module in isolation. For instance, the developer can check his sorting function in isolation, rather than checking in an integrated fashion. The following figure shows the interaction between the entire project test plan.
(B) How Do Test Documents in a Project Span Across the Software Development Lifecycle?
The following figure shows pictorially how test documents span across the software development lifecycle. The following discusses the specific testing documents in the lifecycle: Central/Project test plan: This is the main test plan which outlines the complete test strategy of the software project. This document should be prepared before the start of the project and is used until the end of the software development lifecyle.
Figure 38: Test documents across phases Acceptance test plan: This test plan is normally prepared with the end customer. This document commences during the requirement phase and is completed at final delivery. System test plan: This test plan starts during the design phase and proceeds until the end of the project. Integration and unit test plan: Both of these test plans start during the execution phase and continue until the final delivery. Note The above answer is a different interpretation of V-model testing. We have explained the V-model in this chapter in more detail in one of the questions. Read it once to understand the concept.
(A) Can You Explain Inventories? OR (A) How Do You Do Analysis and Design for Testing Projects? OR (A) Can You Explain Calibration?
The following are three important steps for doing analysis and design for testing: Test objectives: These are broad categories of things which need to be tested in the application. For instance, in the following figure we have four broad categories of test areas: polices, error checking, features, and speed. Inventory: Inventory is a list of things to be tested for an objective. For instance, the following figure shows that we have identified inventory such as add new policy, which is tested for the object types of policies. Change/add address and delete customer is tested for the features objective.
Figure 39: Software testing planning and design Tracking matrix: Once we have identified our inventories we need to map the inventory to test cases. Mapping of inventory to the test cases is called calibration.
Figure 40: Calibration The following is a sample inventory tracking matrix. "Features" is the objective and "add new policy," "change address," and "delete a customer" are the inventory for the objective. Every inventory is mapped to a test case. Only the "delete a customer" inventory is not mapped to any test case. This way we know if we have covered all the aspects of the application in testing. The inventory tracking matrix gives us a quick global view of what is pending and hence helps us to also measure coverage of the application. The following figure shows the "delete a customer" inventory is not covered by any test case thus alerting us of what is not covered.
Figure 41: Inventory tracking matrix Note During the interview try to explain all of the above three steps because that's how testing is planned and designed in big companies. Inventory forms the main backbone of software testing.
(B) Which Test Cases are Written First: White Boxes or Black Boxes?
Normally black box test cases are written first and white box test cases later. In order to write black box test cases we need the requirement document and, design or project plan. All these documents are easily available at the initial start of the project. White box test cases cannot be started in the initial phase of the project because they need more architecture clarity which is not available at the start of the project. So normally white box test cases are written after black box test cases are written. Black box test cases do not require system understanding but white box testing needs more structural understanding. And structural understanding is clearer in the later part of project, i.e., while executing or designing. For black box testing you need to only analyze from the functional perspective which is easily available from a simple requirement document.
Minor: Very low impact but does not affect operations on a large scale.
Major: Affects operations on a very large scale. Critical: Brings the system to a halt and stops the show.
(I) Can You Explain the Waterfall Model? OR (I) Can You Explain the Big-Bang Waterfall Model? OR (I) Can You Explain the Phased Waterfall Model? OR (I) Explain the Iterative Model, Incremental Model, Spiral Model, Evolutionary Model and The V-Model? OR (I) Explain Unit Testing, Integration Tests, System Testing and Acceptance Testing?
Every activity has a lifecycle and the software development process is no exception. Even if you are not aware of the SDLC you are still following it unknowingly. But if a software professional is aware of the SDLC he can execute the project in a controlled fashion. The biggest benefit of this awareness is that developers will not start execution (coding) which can really lead to the project running in an uncontrolled fashion. Second, it helps customer and software professionals to avoid confusion by anticipating the problems and issues before hand. In short, the SDLC defines the various stages in a software lifecycle. But before we try to understand what SDLC is all about we need to get a broader view of the beginning and ending of the SDLC. Any project started if it does not have a start and end then its already in trouble. It's like if you go out for a drive you should know where to start and where to end or else you are moving around endlessly. The figure shows a more global view of the how the SDLS starts and ends. Any project should have entry criteria and exit criteria. For instance, a proper estimation document can be an entry criteria condition. That means if you do not have a proper estimation document in place the project will not start. It can also be more practical. If half payment is not received the project will not start. There can be a list of points which need to be completed before a project starts. Finally, there should be an end to the project which defines when the project will end. For instance, if all the test scenarios given by the end customer are completed the project is finished. In the figure we have the entry criteria as an estimation document and the exit criteria as a signed document by the end client saying the software is delivered.
The following figure shows the typical flow in the SDLC which has six main models. Developers can select a model for their project. Waterfall model Big bang model Phased model Iterative model Spiral model Incremental model
Waterfall Model
Let's have a look at the Waterfall Model which is basically divided into two subtypes: Big Bang waterfall model and the Phased waterfall model. As the name suggests waterfall means flow of water which always goes in one direction so when we say Waterfall model we expect that every phase/stage is frozen.
Big Bang Waterfall Model
The figure shows the Waterfall Big Bang model which has several stages and are described below: Requirement stage: During this stage basic business needs required for the project which are from a user perspective are produced as Word documents with simple points or may be in the form of complicated use case documents.
Design stage: Use case document/requirement document is the input for this stage. Here we decide how to design the project technically and produce a technical document which has a class diagram, pseudo code, etc. Build stage: This stage uses technical documents as input so code can be generated as output at this stage. This is where the actual execution of the project takes place. Test stage: Here, testing is done on the source code produced by the build stage and the final software is given the greenlight. Deliver stage: After succeeding in the test stage the final product/project is finally installed at client end for actual production. This stage is the beginning of the maintenance stage.
In the Waterfall Big Bang model, it is assumed that all stages are frozen which means it's a perfect world. But in actual projects such processes are impractical.
Phased Waterfall Model
In this model the project is divided into small chunks and delivered at intervals by different teams. In short, chunks are developed in parallel by different teams and get integrated in the final project. But the disadvantage of this model is that improper planning may lead to project failure during integration or any mismatch of co-ordination between the team may cause failure.
Iterative Model
The Iterative model was introduced because of problems occuring in the Waterfall model. Now let's take a look at the Iterative model which also has a two subtypes:
Incremental Model
In this model work is divided into chunks like the Phase Waterfall model but the difference is that in the Incremental model one team can work on one or many chunks unlike in the Phase Waterfall model.
Spiral Model
This model uses a series of prototypes which refine our understanding of what we are actually going to deliver. Plans are changed if required per refining of the prototype. So everytime refining of the prototype is done the whole process cycle is repeated.
Evolutionary Model
In the Incremental and Spiral model the main problem is for any changes done in the between the SDLC we need to iterate a whole new cycle. For instance, during the final (deliver) stage, if the customer demands a change we have to iterate the whole cycle again which means we need to update all the previous (requirement, technical documents, source code & test plan) stages. In the Evolutionary model, we divide software into small units which can be delivered earlier to the customer's end. In later stages we evolve the software with new customer needs.
Not e
V-model
This type of model was developed by testers to emphasize the importance of early testing. In this model testers are involved from the requirement stage itself. The following diagram (V-model cycle diagram) shows how for every stage some testing activity is done to ensure that the project is moving forward as planned.
For instance, In the requirement stage we have acceptance test documents created by the testers. Acceptance test documents outline that if these tests pass then the customer will accept the software. In the specification stage testers create the system test document. In the following section, system testing is explained in more detail. In the design stage we have the integration documents created by testers. Integration test documents define testing steps for how the components should work when integrated. For instance, you develop a customer class and product class. You have tested the customer class and the product class individually. But in a practical scenario the customer class will interact with the product class. So you also need to test to ensure the customer class is interacting with the product class properly. In the implement stage we have unit documents created by the programmers or testers.
Starting from the bottom the first test level is "Unit Testing." It involves checking that each feature specified in the "Component Design" has been implemented in the component. In theory, an independent tester should do this, but in practice the developer usually does it, as they are the only people who understand how a component works. The problem with a component is that it performs only a small part of the functionality of a system, and it relies on cooperating with other parts of the system, which may not have been built yet. To overcome this, the developer either builds, or uses, special software to trick the component into believing it is working in a fully functional system.
Integration Testing
As the components are constructed and tested they are linked together to make sure they work with each other. It is a fact that two components that have passed all their tests, when connected to each other, produce one new component full of faults. These tests can be done by specialists, or by the developers. Integration testing is not focused on what the components are doing but on how they communicate with each other, as specified in the "System Design." The "System Design" defines relationships between components. The tests are organized to check all the interfaces, until all the components have been built and interfaced to each other producing the whole system.
System Testing
Once the entire system has been built then it has to be tested against the "System Specification" to see if it delivers the features required. It is still developer focused, although specialist developers known as systems testers are normally employed to do it. In essence, system testing is not about checking the individual parts of the design, but about checking the system as a whole. In fact, it is one giant component. System testing can involve a number of special types of tests used to see if all the functional and non-functional requirements have been met. In addition to functional requirements these may include the following types of testing for the non-functional requirements: Performance - Are the performance criteria met? Volume - Can large volumes of information be handled? Stress - Can peak volumes of information be handled? Documentation - Is the documentation usable for the system? Robustness - Does the system remain stable under adverse circumstances?
There are many others, the need for which is dictated by how the system is supposed to perf(I) What's the Difference between System Testing and Acceptance Testing?
Acceptance testing checks the system against the "Requirements." It is similar to System testing in that the whole system is checked but the important difference is the change in focus: System testing checks that the system that was specified has been delivered. Acceptance testing checks that the system will deliver what was requested. The customer should always do Acceptance testing and not the developer. The customer knows what is required from the system to achieve value in the business and is the only person qualified to make that judgment. This testing is more about ensuring that the software is delivered as defined by the customer. It's like getting a greenlight from the customer that the software meets expectations and is ready to be used.
(B) Can You Explain Boundary Value Analysis? OR (B) What is a Boundary Value in Software Testing?
In some projects there are scenarios where we need to do boundary value testing. For instance, let's say for a bank application you can withdraw a maximum of 25000 and a minimum of 100. So in boundary value testing we only test the exact boundaries rather than hitting in the middle. That means we only test above the max and below the max. This covers all scenarios. The following figure shows the boundary value testing for the bank application which we just described. TC1 and TC2 are sufficient to test all conditions for the bank. TC3 and TC4 are just duplicate/redundant test cases which really do not add any value to the testing. So by applying proper boundary value fundamentals we can avoid duplicate test cases, which do not add value to the testing.
Figure 51: Boundary value analysis So apply the test below to see if it forms an equivalence class or not:
All the test cases should test the same thing. They should produce the same results. If one test case catches a bug, then the other should also catch it. If one of them does not catch the defect, then the other should not catch it.
Figure 52: Equivalence partitioning The following figure shows how the equivalence partition works. Below, we have a scenario in which valid values lie between 20 and 2000.
Figure 53: Sample of equivalence partitioning Any values beyond 2000 and below 20 are invalid. In the following scenario the tester has made four test cases:
Check below 20 (TC1) Check above 2000 (TC2) Check equal to 30 (TC3) Check equal to 1000 (TC4)
Test cases 3 and 4 give the same outputs so they lie in the same partition. In short, we are doing redundant testing. Both TC3 and TC4 fall in one equivalence partitioning, so we can prepare one test case by testing one value in between the boundary, thus eliminating redundancy testing in projects.
(B) Can You Explain How the State Transition Diagrams Can Be Helpful During Testing?
Before we understand how state transition diagrams can be useful in testing, let's understand what exactly a state and transition. The result of a previous input is called a state and transitions are actions which cause the state to change from one state to another. The following figure shows a typical state transition diagram. The arrows signify the transition and the oval shapes signify the states. The first transition in the diagram is the issue of the check that it is ready to be deposited. The second transition is the check is deposited of which we can have two states: either the check cleared or it bounced.
Figure 54: Sample state transition diagram Now that we are clear about state and transition, how does it help us in testing? By using states and transitions we can identify test cases. So we can identify test cases either using states or transitions. But if we use only one entity, i.e., either state or transition, it is very possible that we can miss some scenarios. In order to get the maximum benefit we should use the combination of state and transition. The following figure shows that if we only use state or transition in isolation it's possible that we will have partial testing. But the combination of state and transition can give us better test coverage for an application.
(B) Can You Explain Random Testing? OR (B) Can You Explain Monkey Testing?
Random testing is sometimes called monkey testing. In Random testing, data is generated randomly often using a tool. For instance, the following figure shows how randomlygenerated data is sent to the system. This data is generated either using a tool or some automated mechanism. With this randomly generated input the system is then tested and results are observed accordingly.
Figure 56: Random/Monkey testing Random testing has the following weakness:
They are not realistic. Many of the tests are redundant and unrealistic. You will spend more time analyzing results. You cannot recreate the test if you do not record what data was used for testing.
This kind of testing is really of no use and is normally performed by newcomers. Its best use is to see if the system will hold up under adverse effects.
As the name specifies semi-random testing is nothing but controlling random testing and removing redundant test cases. So what we do is perform random test cases and equivalence partitioning to those test cases, which in turn removes redundant test cases, thus giving us semi-random test cases.
(I) What is an Orthogonal Arrays? OR (I) Can You Explain a Pair-Wise Defect?
Orthogonal array is a two-dimensional array in which if we choose any two columns in the array and all the combinations of numbers will appear in those columns. The following figure shows a simple L9 (34) orthogonal array. In this the number 9 indicates that it has 9 rows. The number 4 indicates that it has 4 columns and 3 indicates that each cell contains a 1, 2, and 3. Choose any two columns. Let's choose column 1 and 2. It has (1,1), (1,2), (1,3), (2,1), (2,2), (2,3), (3,1), (3,2), (3,3) combination values. As you can see these values cover all the values in the array. Compare the values with the combination of column 3 and 4 and they will fall in some pair. This is applied in software testing which helps us eliminate duplicate test cases.
Figure 60: Sample orthogonal array Now let's try to apply an orthogonal array in actual testing field. Let's say we have a scenario in which we need to test a mobile handset with different plan types, terms, and sizes. Below are the different situations:
Handset (Nokia, 3G and Orange). Plan type (4 x 400, 4 x 300, and 2 x 270). Term (Long-term, short-term, and mid-term). Size (3, 4, and 5 inch).
Each handset should be tested with every plan type, term, and size. Each plan type should be tested with every handset, term, and size. Each size should be tested with every handset, plan type, and term.
So now you must be thinking we have 81 combinations. But we can test all these conditions with only 9 test cases. The following is the orthogonal array for it.
Figure 61: Orthogonal array in actual testing Orthogonal arrays are very useful because most defects are pairs-wise defects and with orthogonal arrays, we can reduce redundancy to a huge extent.
Figure 62: General decision tables The following is a sample decision table for a discount which depends on age. Discounts are only allowed if you are married or a student. The following is the decision table accordingly. Using the decision table we have also derived our test cases. Because this is a sample example we cannot see the importance of the decision table. But just imagine that you have a huge amount of possible inputs and outputs. For such a scenario decision tables gives you a better view.
Figure 63: Discount Decision table The following is the decision table for the scenarios described above. In the top part we have put the condition and below are the actions which occur as a result of the conditions. Read from the right and move to the left and then to the action. For instance, Married Yes then discount. The same goes for the student condition. Using the decision table we can ensure, to a good extent, that we do not skip any validation in a project.
Severity 1 (showstoppers): These kinds of defects do not allow the application to move ahead. So they are also called showstopper defects. Severity 2 (application continues with severe defects): Application continues working with these types of defects, but they can have high implications, later, which can be more difficult to remove. Severity 3 (application continues with unexpected results): In this scenario the application continues but with unexpected results. Severity 4 (suggestions): Defects with these severities are suggestions given by the customer to make the application better. These kinds of defects have the least priority and are considered at the end of the project or during the maintenance stage of the project.
Identify the risk of the project by discussion, proper requirement gathering, and forecasting. Once you have identified the risk prioritize which risk has the most impact and should be tackled on a priority basis. Analyze how the risk can be solved by proper impact analysis and planning. Finally, using the above analysis, we mitigate the risk.
(I) What are the Different Cost Elements Involved in Implementing a Process in an Organization?
Below are some of the cost elements involved in the implementing process:
Salary: This forms the major component of implementing any process, the salary of the employees. Normally while implementing a process in a company either organization can recruit full-time people or they can share resources part-time for implementing the process. Consultant: If the process is new it can also involve consultants which is again an added cost. Training Costs: Employees of the company may also have to undergo training in order to implement the new process. Tools: In order to implement the process an organization will also need to buy tools which again need to be budgeted for. Implementing a process is not an easy job in any organization. More than financial commitment, it requires commitment from people to follow the process.
Note
Figure 68: Model Many companies reinvent the wheel rather than following time tested models in the industry.
Chapter 4: CMMI
(B) What is CMMI and What's the Advantage of Implementing it in an Organization?
CMMI stands for Capability Maturity Model Integration. It is a process improvement approach that provides companies with the essential elements of an effective process. CMMI can serve as a good guide for process improvement across a project, organization, or division. CMMI was formed by using multiple previous CMM processes. The following are the areas which CMMI addresses:
Systems engineering: This covers development of total systems. System engineers concentrate on converting customer needs to product solutions and supports them throughout the product lifecycle. Software engineering: Software engineers concentrate on the application of systematic, disciplined, and quantifiable approaches to the development, operation, and maintenance of software. Integrated Product and Process Development (IPPD): Integrated Product and Process Development (IPPD) is a systematic approach that achieves a timely collaboration of relevant stakeholders throughout the life of the product to better satisfy customer needs, expectations, and requirements. This section mostly concentrates on the integration part of the project for different processes. For instance, it's possible that your project is using services of some other third party component. In such situations the integration is a big task itself, and if approached in a systematic manner, can be handled with ease. Software acquisition: Many times an organization has to acquire products from other organizations. Acquisition is itself a big step for any organization and if not handled in a proper manner means a disaster is sure to happen. The following figure shows the areas involved in CMMI.
(I) What are Different Models in CMMI? OR (I) Can You Explain Staged and Continuous Models in CMMI?
There are two models in CMMI. The first is "staged" in which the maturity level organizes the process areas. The second is "continuous" in which the capability level organizes the process area.
Figure 75: CMMI models The following figure shows how process areas are grouped in both models.
(I) Can You Explain the Different Maturity Levels in a Staged Representation?
There are five maturity levels in a staged representation as shown in the following figure. Maturity Level 1 (Initial): In this level everything is adhoc. Development is completely chaotic with budget and schedules often exceeded. In this scenario we can never predict quality. Maturity Level 2 (Managed): In the managed level basic project management is in place. But the basic project management and practices are followed only in the project level.
Maturity Level 3 (Defined): To reach this level the organization should have already achieved level 2. In the previous level the good practices and process were only done at the project level. But in this level all these good practices and processes are brought to the organization level. There are set and standard practices defined at the organization level which every project should follow. Maturity Level 3 moves ahead with defining a strong, meaningful, organizational approach to developing products. An important distinction between Maturity Levels 2 and 3 is that at Level 3, processes are described in more detail and more rigorously than at Level 2 and are at an organization level. Maturity Level 4 (Quantitively measured): To start with, this level of organization should have already achieved Level 2 and Level 3. In this level, more statistics come into the picture. Organization controls the project by statistical and other quantitative techniques. Product quality, process performance, and service quality are understood in statistical terms and are managed throughout the life of the processes. Maturity Level 4 concentrates on using metrics to make decisions and to truly measure whether progress is happening and the product is becoming better. The main difference between Levels 3 and 4 are that at Level 3, processes are qualitatively predictable. At Level 4, processes are quantitatively predictable. Level 4 addresses causes of process variation and takes corrective action. Maturity Level 5 (Optimized): The organization has achieved goals of maturity levels 2, 3, and 4. In this level, processes are continually improved based on an understanding of common causes of variation within the processes. This is like the final level; everyone on the team is a productive member, defects are minimized, and products are delivered on time and within the budget boundary. The following figure shows, in detail, all the maturity levels in a pictorial fashion.
This level means that any generic or specific practice of capability level 1 is not performed.
Capability Level 1: Performed
The capability level 1 process is expected to perform all capability level 1 specific and generic practices for that process area. In this level performance may not be stable and probably does not meet objectives such as quality, cost, and schedule, but still the task can be done.
Capability Level 2: Managed
Capability level 2 is a managed process planned properly, performed, monitored, and controlled to achieve a given purpose. Because the process is managed we achieve other objectives, such as cost, schedule, and quality. Because you are managing, certain metrics are consistently collected and applied to your management approach.
Capability Level 3: Defined
The defined process is a managed process that is tailored from an organization standard. Tailoring is done by justification and documentation guidelines. For instance your organization may have a standard that we should get an invoice from every supplier. But if the supplier is not able to supply the invoice then he should sign an agreement in place of the invoice. So here the invoice standard is not followed but the deviation is under control.
Capability Level 4: Quantitatively Managed
The quantitatively managed process is a defined process which is controlled through statistical and quantitative information. So from defect tracking to project schedules all are statistically tracked and measured for that process.
The optimizing process is a quantitatively managed process where we increase process performance through incremental and innovative improvements. Continuous representation is the same as staged only that information is arranged in a different fashion. The biggest difference is one concentrates on a specific process while the other brings a group of processes to a certain maturity level.
(A) How Many Process Areas are Present in CMMI and What Classification Do They Fall in?
All 25 process areas in CMMI are classified into four major sections.
Process Management
This process areas contain all project tasks related to defining, planning, executing, implementing, monitoring, controlling, measuring, and making better processes.
Project Management
Project management process areas cover the project management activities related to planning, monitoring, and controlling the project.
Engineering
Engineering process areas were written using general engineering terminology so that any technical discipline involved in the product development process (e.g., software engineering or mechanical engineering) can use them for process improvement.
Support
Support process areas address processes that are used in the context of performing other processes. In general, the support process areas address processes that are targeted toward the project and may address processes that apply more generally to the organization. For example, process and product quality assurance can be used with all the process areas to provide an objective evaluation of the processes and work products described in all the process areas. The following diagram shows the classification and representation of the process areas.
The following table defines all the abbreviations of the process areas.
These levels are the biggest steps for any organization because the organization moves from a immature position to a more mature organization. Level 1 is an adhoc process in which people have created a personal process to accomplish a certain task. With this approach there is a lot of redundant work and people do not share their information. This leads to heroes' in the project, so when people move out of the organization the knowledge also moves out, and the organization suffers. In maturity level 2 individuals share their lessons and best practices, which leads to devising preliminary processes at the project and in some cases it also moves to the organization level. In level 2 we focus on project management issues that affect day to day routines. It has seven process areas as shown in the figure below. So in the short difference between level 1 and level 2 is related to immature and mature organizations.
Now that in Level 2 good practices are observed at the project level, it is time to move these good practices to the organization level so that every one can benefit from the same. So the biggest difference between Level 2 and Level 3 is good practices from the projects that are bubbled up to organization level. The organization approach of doing business is documented. To perform Maturity level 3, first Maturity 2 must be achieved with the 14 processes as shown in the given figure.
Maturity level 4 is all about numbers and statistics. All aspects of the project are managed by numbers. All decisions are made by numbers. Product quality and process are measured by numbers. So in Level 3 we say this is of good quality; in Level 4 we say this is of good quality because the defect ratio is less than 1 %. So there are two process areas in Level 4 as shown below. In order to move to Level 4, you should have achieved all the PA's of Level 3 and also the two process areas below.
Level 5 is all about improvement as compared to Level 4. Level 5 concentrates on improving quality of organization process by identifying variation, by looking at root causes of the conditions and incorporating improvements for improve process. Below are the two process areas in Level 5 as shown in figure below. In order to get level 5 all level 4 PA's should be satisfied. So the basic difference between level 4 and level 5 is in Level 4 we have already achieved a good level of quality, and in level 5 we are trying to improve the quality.
(I) What Different Sources are Needed to Verify Authenticity for CMMI Implementation?
There are three different sources from which an appraiser can verify that an organization followed the process or not. Instruments: An instrument is a survey or questionnaire provided to the organization, project, or individuals before starting the assessment so that beforehand the appraiser knows some basic details of the project.
Interview: An interview is a formal meeting between one or more members of the organization in which they are asked some questions and the appraiser makes some judgments based on those interviews. During the interview the member represents some process area or role which he performs. For instance, the appraiser may interview a tester or programmer asking him indirectly what metrics he has submitted to his project manager. By this the appraiser gets a fair idea of CMMI implementation in that organization. Documents: A document is a written work or product which serves as evidence that a process is followed. It can be hard copy, Word document, email, or any type of written official proof. The following figure is the pictorial view of the sources used to verify how compliant the organization is with CMMI.
(I) Can You Explain the SCAMPI Process? OR (I) How is Appraisal Done in CMMI?
SCAMPI stands for Standard CMMI Appraisal Method for Process Improvement. SCAMPI is an assessment process used to get CMMI certified for an organization. There are three classes of CMMI appraisal methods: Class A, Class B, and Class C. Class A is the most aggressive, while Class B is less aggressive, and Class C is the least aggressive.
Figure 88: SCAMPI Let's discuss these appraisal methods in more detail. Class A: This is the only method that can provide a rating and get you a CMMI certificate. It requires all three sources of data instruments, interviews, and documents.
Class B: This class requires only two sources of data (interviews and either documents or instruments). But please note you do not get rated with Class B appraisals. Class B is just a warm-up to see if an organization is ready for Class A. With less verification the appraisal takes less time. In this class data sufficiency and draft presentations are optional. Class C: This class requires only one source of data (interviews, instruments, or documents). Team consensus, validation, observation, data sufficiency, and draft presentation are optional. The following table shows the characteristic features with proper comparison.
Use Class B to initiate a process improvement plan. After that apply Class C to check readiness for Class B or Class A. The following diagram shows this strategy.
Second Strategy
Class C appraisal is used on a subset of an organization. From this we get an aggregation of weakness across the organization. From this we can prepare a process improvement plan. We can then apply a Class B appraisal to see if we are ready for Class A appraisal. The following diagram shows the strategy.
Class A is used to initiate an organization level process. The process improvement plan is based on an identified weakness. Class B appraisal should be performed after six months to see the readiness for the second Class A appraisal rating. The following diagram shows this strategy.
Figure 93: Sample PIID Once the PII documents are filed we can rate whether the organization is compliant or not. Below are the steps to be followed during the SCAMPI:
(A) Can You Explain Implementation of CMMI in One of the Key Process Areas?
Note This question will be asked to judge whether you have actually implemented CMMI in a proper fashion in your oganization. To answer this question, we will be using SAM as the process area. But you can answer with whatever process area you have implemented in your organization.
For the following SAM process there are the two SG1 and SG2 practices which need to be implemented to satisfy the process area. SAM helps us to define our agreement with the supplier while procuring products in the company. Let's see, in the next step, how we have mapped our existing process with the SAM practices defined in CMMI.
SAM is a process adopted by the company. If anyone wants to demand any product he has to first raise demand for the item by using the demand form which is defined by the company. Depending on demand the supervisor defines which acquisition type the demand is. For instance, is it a production acquisition type, office material acquisition type, or another type. Once the acquisition type is decided the organization places an advertisement in the newspaper to ask suppliers for quotes. Once all quotations are received depending on cost, quality, and other factors the final supplier is decided. The supplier is then called to the office and he has to sign an agreement with the organization for the delivery of the product. Once the agreement is signed the supplier sends a sample product which is analyzed by the organization practically. Finally, the product is accepted and the supplier is then asked to send the complete delivery of all products. The product is accepted in the organization by issuing the supplier a proper invoice. The invoice document says that the product is accepted by
the organization officially. When the product is installed in the organization then either someone from the supplier side comes for the demo or a help brochure is shipped with the product.
The above explanation is from the perspective of the how the organization manages its transactions with the supplier. Now let's try to map how the above process fits in the CMMI model. In the above diagram the circled descriptions are process areas of CMMI. Open table as spreadsheet Organization process In the above process the demand form decides what the acquisition type of the product is. Looking at the quotation the supplier is reviewed and the selection is done.
Establish supplier In the above process we have a step when we sign an agreement with agreements the supplier which establishes all the terms and conditions for the supplier agreement. Review product One of the steps of the process is that the supplier has to send a sample which is reviewed by the organization. The supplier agreement is executed by accepting the invoice.
In the above process the transition of the product either happens through help brochures or when the demo person visits.
(B) What are All the Process Areas and Goals and Practices? OR (A) Can You Explain All the Process Areas?
Not e
No one is going to ask such a question, but they would like to know at least the purpose of each KPA. Second, they would like to know what you did to attain compatibility in these process areas. For instance, you say that you did an organizational process. They would like to know how you did it. You can justify it by saying that you made standard documents for coding standards which was then followed at the organization level for reference. Normally everyone follows a process; only they do not realize it. So try to map the KPA to the process that you followed.
Each process area is defined by a set of goals and practices. There are two categories of goals and practices: generic and specific. Generic goals and practices are a part of every process area. Specific goals and practices are specific to a given process area. A process area is satisfied when company processes cover all of the generic and specific goals and practices for that process area. Generic goals and practices are a part of every process area. They include the following: GG 1 GP 1.1 GG 2 GP 2.1 GP 2.2 GP 2.3 GP 2.4 Achieve Specific Goals Perform Base Practices Institutionalize a Managed Process Establish an Organizational Policy Plan the Process Provide Resources Assign Responsibility
GP 2.5 GP 2.6 GP 2.7 GP 2.8 GP 2.9 GP 2.10 GG 3 GP 3.1 GP 3.2 GG 4 GP 4.1 GP 4.2 GG 5 GP 5.1 GP 5.2
Train People Manage Configurations Identify and Involve Relevant Stakeholders Monitor and Control the Process Objectively Evaluate Adherence Review Status with Higher Level Management Institutionalize a Defined Process Establish a Defined Process Collect Improvement Information Institutionalize a Quantitatively Managed Process Establish Quantitative Objectives for the Process Stabilize Sub process Performance Institutionalize an Optimizing Process Ensure Continuous Process Improvement Correct Root Causes of Problems
Process Areas
The CMMI contains 25 key process areas indicating the aspects of product development that are to be covered by company processes.
Causal Analysis and Resolution (CAR)
Purpose
The purpose of Causal Analysis and Resolution (CAR) is to identify causes of defects and other problems and take action to prevent them from occurring in the future.
Specific Practices by Goal
Determine Causes of Defects Select Defect Data for Analysis Analyze Causes Address Causes of Defects Implement the Action Proposals Evaluate the Effect of Changes Record Data
The purpose of Configuration Management (CM) is to establish and maintain the integrity of work products using configuration identification, configuration control, configuration status accounting, and configuration audits.
Specific Practices by Goal
Establish Baselines Identify Configuration Items Establish a Configuration Management System Create or Release Baselines Track and Control Changes
Track Change Requests Control Configuration Items Establish Integrity Establish Configuration Management Records Perform Configuration Audits
The purpose of Decision Analysis and Resolution (DAR) is to analyze possible decisions using a formal evaluation process that evaluates identified alternatives against established criteria.
Specific Practices by Goal
Evaluate Alternatives Establish Guidelines for Decision Analysis Establish Evaluation Criteria Identify Alternative Solutions Select Evaluation Methods Evaluate Alternatives Select Solutions
Purpose
The purpose of Integrated Project Management (IPM) is to establish and manage the project and the involvement of the relevant stakeholders according to an integrated and defined process that is tailored from the organization's set of standard processes.
Specific Practices by Goal
SG 1 SP 1.1-1 SP 1.2-1 SP 1.3-1 SP 1.4-1 SP 1.5-1 SG 2 SP 2.1-1 SP 2.2-1 SP 2.3-1 SG 3 SP 3.1-1 SP 3.2-1 SG 4 SP 4.1-1 SP 4.2-1 SP 4.3-1
Use the Project's Defined Process Establish the Project's Defined Process Use Organizational Process Assets for Planning Project Activities Integrate Plans Manage the Project Using the Integrated Plans Contribute to the Organizational Process Assets Coordinate and Collaborate with Relevant Stakeholders Manage Stakeholder Involvement Manage Dependencies Resolve Coordination Issues Use the Project's Shared Vision for IPPD Define a Project's Shared Vision for IPPD Establish the Project's Shared Vision Organize Integrated Teams for IPPD Determine Integrated Team Structure for the Project Develop a Preliminary Distribution of Requirements to Integrated Teams Establish Integrated Teams
The purpose of Integrated Supplier Management (ISM) is to proactively identify sources of products that may be used to satisfy the project's requirements and to manage selected suppliers while maintaining a cooperative project-supplier relationship.
Specific Practices by Goal
Analyze and Select Sources of Products Analyze Potential Sources of Products Evaluate and Determine Sources of Products Coordinate Work with Suppliers Monitor Selected Supplier Processes Evaluate Selected Supplier Work Products Revise the Supplier Agreement or Relationship
The purpose of Integrated Teaming (IT) is to form and sustain an integrated team for the development of work products.
Specific Practices by Goal
SG 1 SP 1.1-1 SP 1.2-1
Establish Team Composition Identify Team Tasks Identify Needed Knowledge and Skills
Assign Appropriate Team Members Govern Team Operation Establish a Shared Vision Establish a Team Charter Define Roles and Responsibilities Establish Operating Procedures Collaborate among Interfacing Teams
The purpose of Measurement and Analysis (MA) is to develop and sustain a measurement capability that is used to support management information needs.
Specific Practices by Goal
Align Measurement and Analysis Activities Establish Measurement Objectives Specify Measures Specify Data Collection and Storage Procedures Specify Analysis Procedures Provide Measurement Results Collect Measurement Data Analyze Measurement Data Store Data and Results
SP 2.4-1
Communicate Results
The purpose of Organizational Environment for Integration (OEI) is to provide an Integrated Product and Process Development (IPPD) infrastructure and manage people for integration.
Specific Practices by Goal
Provide IPPD Infrastructure Establish the Organization's Shared Vision Establish an Integrated Work Environment Identify IPPD-Unique Skill Requirements Manage People for Integration Establish Leadership Mechanisms Establish Incentives for Integration Establish Mechanisms to Balance Team and Home Organization Responsibilities
The purpose of Organizational Innovation and Deployment (OID) is to select and deploy incremental and innovative improvements that measurably improve the organization's processes and technologies. The improvements support the organization's quality and process-performance objectives as derived from the organization's business objectives.
Select Improvements Collect and Analyze Improvement Proposals Identify and Analyze Innovations Pilot Improvements Select Improvements for Deployment Deploy Improvements Plan the Deployment Areas Manage the Deployment Measure Improvement Effects
The purpose of the Organizational Process Definition (OPD) is to establish and maintain a usable set of organizational process assets.
Specific Practices by Goal
Establish Organizational Process Assets Establish Standard Processes Establish Life-Cycle Model Descriptions Establish Tailoring Criteria and Guidelines Establish the Organization's Measurement Repository Establish the Organization's Process Asset Library
The purpose of Organizational Process Focus (OPF) is to plan and implement organizational process improvement based on a thorough understanding of the current strengths and weaknesses of the organization's processes and process assets.
Specific Practices by Goal
Determine Process Improvement Opportunities Establish Organizational Process Needs Appraise the Organization's Processes Identify the Organization's Process Improvements Plan and Implement Process Improvement Activities Establish Process Action Plans Implement Process Action Plans Deploy Organizational Process Assets Incorporate Process-Related Experiences into the Organizational Process Assets
The purpose of Organizational Process Performance (OPP) is to establish and maintain a quantitative understanding of the performance of the organization's set of standard processes in support of quality and process-performance objectives, and to provide the process performance data, baselines, and models to quantitatively manage the organization's projects.
Specific Practices by Goal
SG 1
Select Processes Establish Process Performance Measures Establish Quality and Process Performance Objectives Establish Process Performance Baselines Establish Process Performance Models
The purpose of Organizational Training (OT) is to develop the skills and knowledge of people so that they can perform their roles effectively and efficiently.
Specific Practices by Goal
Establish an Organizational Training Capability Establish the Strategic Training Needs Determine Which Training Needs Are the Responsibility of the Organization Establish an Organizational Training Tactical Plan Establish Training Capability Provide Necessary Training Deliver Training Establish Training Records Assess Training Effectiveness
The purpose of Product Integration (PI) is to assemble the product from the product components, ensure that the product is integrated, functions properly, and delivers the product.
Specific Practices by Goal
Prepare for Product Integration Determine Integration Sequence Establish the Product Integration Environment Establish Product Integration Procedures and Criteria Ensure Interface Compatibility Review Interface Descriptions for Completeness Manage Interfaces Assemble Product Components and Deliver the Product Confirm Readiness of Product Components for Integration Assemble Product Components Evaluate Assembled Product Components Package and Deliver the Product or Product Component
Purpose
The purpose of Project Monitoring and Control (PMC) is to provide an understanding of the project's progress so that appropriate corrective actions can be taken when the project's performance deviates significantly from the plan.
Specific Practices by Goals
SG 1 SP 1.1-1 SP 1.2-1 SP 1.3-1 SP 1.4-1 SP 1.5-1 SP 1.6-1 SP 1.7-1 SG 2 SP 2.1-1 SP 2.2-1 SP 2.3-1
Monitor Project Against the Plan Monitor Project Planning Parameters Monitor Commitments Monitor Project Risks Monitor Data Management Monitor Stakeholder Involvement Conduct Progress Reviews Conduct Milestone Reviews Manage Corrective Action to Closure Analyze Issues Take Corrective Action Manage Corrective Action
The purpose of Project Planning (PP) is to establish and maintain plans that define project activities.
SG 1 SP 1.1-1 SP 1.2-1 SP 1.3-1 SP 1.4-1 SG 2 SP 2.1-1 SP 2.2-1 SP 2.3-1 SP 2.4-1 SP 2.5-1 SP 2.6-1 SP 2.7-1 SG 3 SP 3.1-1 SP 3.2-1 SP 3.3-1
Establish Estimates Estimate the Scope of the Project Establish Estimates of Work Product and Task Attributes Define Project Life Cycle Determine Estimates of Effort and Cost Develop a Project Plan Establish the Budget and Schedule Identify Project Risks Plan for Data Management Plan for Project Resources Plan for Needed Knowledge and Skills Plan Stakeholder Involvement Establish the Project Plan Obtain Commitment to the Plan Review Plans that Affect the Project Reconcile Work and Resource Levels Obtain Plan Commitment
Purpose
The purpose of Process and Product Quality Assurance (PPQA) is to provide staff and management with objective insight into processes and associated work products.
Specific Practices by Goal
Objectively Evaluate Processes and Work Products Objectively Evaluate Processes Objectively Evaluate Work Products and Services Provide Objective Insight Communicate and Ensure Resolution of Noncompliance Issues Establish Records
The purpose of the Quantitative Project Management (QPM) process area is to quantitatively manage the project's defined process to achieve the project's established quality and process-performance objectives.
Specific Practices by Goal
Quantitatively Manage the Project Establish the Project's Objectives Compose the Defined Processes Select the Sub processes that Will Be Statistically Managed Manage Project Performance Statistically Manage Sub-process Performance
Select Measures and Analytic Techniques Apply Statistical Methods to Understand Variation Monitor Performance of the Selected Sub-processes Record Statistical Management Data
The purpose of Requirements Development (RD) is to produce and analyze customer, product, and product-component requirements.
Specific Practices by Goal
Develop Customer Requirements Collect Stakeholder Needs Elicit Needs Develop the Customer Requirements Develop Product Requirements Establish Product and Product-Component Requirements Allocate Product-Component Requirements Identify Interface Requirements Analyze and Validate Requirements Establish Operational Concepts and Scenarios Establish a Definition of Required Functionality Analyze Requirements
Analyze Requirements to Achieve Balance Validate Requirements Validate Requirements with Comprehensive Methods
The purpose of Requirements Management (REQM) is to manage the requirements of the project's products and product components and to identify inconsistencies between those requirements and the project's plans and work products.
Specific Practices by Goal
Manage Requirements Obtain an Understanding of Requirements Obtain Commitment to Requirements Manage Requirements Changes Maintain Bidirectional Traceability of Requirements Identify Inconsistencies between Project Work and Requirements
The purpose of Risk Management (RSKM) is to identify potential problems before they occur so that risk-handling activities can be planned and invoked as needed across the life of the product or project to mitigate adverse impacts on achieving objectives.
Prepare for Risk Management Determine Risk Sources and Categories Define Risk Parameters Establish a Risk Management Strategy Identify and Analyze Risks Identify Risks Evaluate, Categorize, and Prioritize Risks Mitigate Risks Develop Risk Mitigation Plans Implement Risk Mitigation Plans
The purpose of the Supplier Agreement Management (SAM) is to manage the acquisition of products from suppliers for which there exists a formal agreement.
Specific Practices by Goal
Establish Supplier Agreements Determine Acquisition Type Select Suppliers Establish Supplier Agreements Satisfy Supplier Agreements
Review COTS Products Execute the Supplier Agreement Accept the Acquired Product Transition Products
The purpose of the Technical Solution (TS) is to design, develop, and implement solutions to requirements. Solutions, designs, and implementations encompass products, product components, and product-related life-cycle processes either alone or in appropriate combinations.
Specific Practices by Goal
Select Product-Component Solutions Develop Alternative Solutions and Selection Criteria Develop Detailed Alternative Solutions and Selection Criteria Evolve Operational Concepts and Scenarios Select Product-Component Solutions Develop the Design Design the Product or Product Component Establish a Technical Data Package Establish Interface Descriptions Design Interfaces Using Criteria Perform Make, Buy, or Reuse Analyses
SG 3 SP 3.1-1 SP 3.2-1
Implement the Product Design Implement the Design Develop Product Support
The purpose of Validation (VAL) is to demonstrate that a product or product component fulfills its intended use when placed in its intended environment.
Specific Practices by Goal
Prepare for Validation Select Products for Validation Establish the Validation Environment Establish Validation Procedures and Criteria Validate Product or Product Components Perform Validation Analyze Validation Results
Verification (VER)
The purpose of Verification (VER) is to ensure that a selected work product meets their specified requirements.
Specific Practices by Goal
SG 1
Select Work Products for Verification Establish the Verification Environment Establish Verification Procedures and Criteria Perform Peer Reviews Prepare for Peer Reviews Conduct Peer Reviews Analyze Peer Review Data Verify Selected Work Products Perform Verification Analyze Verification Results and Identify Corrective Action.
(I) Can You Explain the Different Methodology for the Execution and the Design Process Stages in Six Sigma?
The main focus of Six Sigma is to reduce defects and variations in the processes. DMAIC and DMADV are the models used in most Six Sigma initiatives. DMADV is the model for designing processes while DMAIC is used for improving the process.
Figure 97: Methodology in Six Sigma The DMADV model includes the following five steps:
Define: Determine the project goals and the requirements of customers (external and internal). Measure: Assess customer needs and specifications. Analyze: Examine process options to meet customer requirements. Design: Develop the process to meet the customer requirements. Verify: Check the design to ensure that it's meeting customer requirements
Define the projects, goals, and deliverables to customers (internal and external). Describe and quantify both the defects and the expected improvements. Measure the current performance of the process. Validate data to make sure it is credible and set the baselines. Analyze and determine the root cause(s) of the defects. Narrow the causal factors to the vital few. Improve the process to eliminate defects. Optimize the vital few and their interrelationships. Control the performance of the process. Lock down the gains.
(I) What are Executive Leaders, Champions, Master Black Belts, Green Belts, and Black Belts?
Six Sigma is not only about techniques, tools, and statistics, but also about people. In Six Sigma there are five key players:
Executive leaders Champions Master black belts Black belts Green belts
Let's try to understand the role of the players step by step. Executive leaders: They are the main people who actually decide that we need to do Six Sigma. They promote it throughout the organization and ensure commitment of the organization. Executive leaders are mainly either CEOs or from the board of directors. So, in short, they are the people who fund the Six Sigma initiative. They should believe that Six Sigma will improve the organization process and that they will succeed. They should ensure that resources get proper training on Six Sigma, understand how it will benefit the organization, and track the metrics. Champions: Champions are normally senior managers of the company. These people promote Six Sigma mainly between the business users. The champion understands Six Sigma thoroughly, and serve as a coaches and mentors, select the project, decide objectives, dedicate resources to black belts, and remove obstacles which come across black belt players. Historically, champions always fight for a cause. In Six Sigma they fight to remove black belt hurdles.
Master black belts: This role requires the highest level of technical capability in Six Sigma. Normally organizations that are just starting up with Six Sigma will not have master black belts. So normally outsiders are recruited. The main role of the master black belt is to train, mentor, and guide. He helps the executive leaders in selecting candidates, finding the right project, teaching the basics, and training resources. They regularly meet with black belts and green belts training and mentoring them. Black belts: Black belts lead a team on a selected project which has to be showcased for Six Sigma. They are mainly responsible for finding out variations and seeing how these variations can be minimized. Master black belts basically select a project and train resources, but black belts are the people who actually implement them. Black belts normally work in projects as team leaders or project managers. They are central to Six Sigma as they are actually implementing Six Sigma in the organization. Green belts: Green belts assist black belts in their functional areas. They are mainly in projects and work part-time on Six Sigma implementation. They apply Six Sigma methodologies to solve problems and improve processes at the bottom level. They have just enough knowledge of Six Sigma and they help to define the base of Six Sigma implementation in the organization. They assist black belts in Six Sigma implementation.
(I) What are the Different Kinds of Variations Used in Six Sigma?
Variation is the basis of Six Sigma. It defines how many changes are happening in the output of a process. So if a process is improved then this should reduce variations. In Six Sigma we identify variations in the process, control them, and reduce or eliminate defects. Now let's discuss how we can measure variations.
Figure 100: Different variations in Six Sigma There are four basic ways of measuring variations: Mean, Median, Mode, and Range. Let's discuss each of these variations in more depth for better analysis. Mean: In the mean measurement the variations are measured and compared using averaging techniques. For instance, you can see from the following figures which shows two weekly measures, how many computers are manufactured. We have tracked two weeks; one we have named Week 1 and the other Week 2. So to calculate variation using mean we calculate the mean of Week 1 and Week 2. You can see from the calculations in the following figure we have 5.083 for Week 1 and 2.85 for Week 2. So we have a variation of 2.23.
Figure 101: Measuring variations using mean Median: Median value is a mid-point in our range of data. The mid-point can be found by finding the difference between the highest and lowest value and then dividing it by two and, finally, adding the lowest value to it. For instance, for the following figure in Week 1 we have 4 as the lowest value and 7 as the highest value. So first we subtract the lowest value from the highest value, i.e., 7 - 4. Then we divide it by two and add the lowest value. So for Week 1 the median is 5.5 and for Week 2 the median is 2.9. So the variation is 5.5-2.9.
Figure 102: Median for calculating variations Range: Range is nothing but the spread of values for a particular data range. In short, it is the difference between the highest and lowest values in a particular data range. For instance, you can see for the recorded computer data of Week 2 we have found the range of values by subtracting the highest value from the lowest.
Figure 103: Range for calculating variations Mode: Mode is nothing but the most frequently occurring values in a data range. For instance, in our computer manufacturing data, range 4 is the most occurring value in Week 1 and 3 is the most occurring value in Week 2. So the variation is 1 between these data ranges.
The most accurate method of quantifying variation is by using standard deviation. It indicates the degree of variation in a set of measurements or a process by measuring the average spread of data around the mean. It's more complicated than the deviation process discussed in the previous question, but it does give accurate information. Below is the formula for standard deviation. The "" symbol stands for standard deviation. X is the observed values; X (with the top bar) is the arithmetic mean; and n is the number of observations. The formula must look complicated but let's break-up into steps to understand it better.
Figure 105: Standard deviation formula The first step is to calculate the mean. This can be calculated by adding up all the observed values and dividing them by the number of observed values. The second step is to subtract the average from each observation, square them, and then sum them. Because we square them we will not get negative values. The following figure shows this very detailed manner.
Figure 107: Step 2Standard deviation In the third step we divide the average by the number of observations as shown in the figure.
Figure 108: Step 3Standard deviation In the final step we take the square root which gives the standard deviation.
Figure 110: Fish bone/Ishikawa diagram The main bone is the problem which we need to address to know what caused the failure. For instance, the following fish bone is constructed to find what caused the project failure. To know this cause we have taken four main bones as inputs: Finance, Process, People, and Tools. For instance, on the people front there are many resignations this was caused because there was no job satisfaction this was caused because the project was a maintenance project. In the same way causes are analyzed on the Tools front also. In Tools No tools were used in the project because no resource had enough knowledge of them this happened because of a lack of planning. In the Process front the process was adhoc this was because of tight deadlines this was caused because marketing people over promised and did not negotiate properly with the end customer. Now once the diagram is drawn the end bones of the fish bone signify the main cause of project failure. From the following diagram here's a list of causes:
No training was provided for the resources regarding tools. Marketing people over promised the customer which lead to tight deadlines. Resources resigned because it's a maintenance project.
Chapter 6: Metrics
(I) Can You Explain How the Number of Defects are Measured?
The number of defects is one of the measures used to measure test effectiveness. One of the side effects of the number of defects is that all bugs are not equal. So it becomes necessary to weight bugs according to there criticality level. If we are using the number of defects as the metric measurement the following are the issues:
The number of bugs that originally existed significantly impacts the number of bugs discovered, which in turns gives a wrong measure of the software quality. All defects are not equal so defects should be numbered with a criticality level to get the right software quality measure.
The following are three simple tables which show the number of defects SDLC phase-wise, module-wise and developer-wise.
(I) Can You Explain How the Number of Production Defects are Measured?
This is one of the most effective measures. The number of defects found in a production is recorded. The only issue with this measure is it can have latent and masked defects which can give us the wrong value regarding software quality.
Figure 115: Defect seeding Let's discuss the concept of defect seeding by doing some detailed calculations and also try to understand how we can predict the number of defects remaining in a system. The following is the calculation used:
1. 2. 3.
First, calculate the seed ratio using the following formula, i.e., number of seed bugs found divided by the total number of seeded bugs. After that we need to calculate the total number of defects by using the formula (number of defects divided by the seed ratio). Finally, we can know the estimated defects by using the formula (total number of defectsthe number of defect calculated by Step 3).
The following figure shows a sample with the step-by-step calculation. You can see that first we calculate the seed ratio, then the total number of defects, and finally, we get the estimated defects.
Figure 117: DRE formula But the success of DRE depends on several factors. The following are some of them:
Severity and distribution of bugs must be taken into account. Second, how do we confirm when the customer has found all the bugs. This is normally achieved by looking at the history of the customer.
DRE is also useful to measure the effectiveness of a particular test such as acceptance, unit, or system testing. The following figure shows defect numbers at various software cycle levels. The + indicates that defects are input at the phase and indicates that these many defects were removed from that particular phase. For instance, in the requirement phase 100 defects were present, but 20 defects are removed from the requirement phase due to a code review. So if 20 defects are removed then 80 defects get carried to the new phase (design) and so on.
First, let's calculate simple DRE of the above diagram. DRE will be the total bugs found in testing divided by the total bugs found in testing plus the total bugs found by the user, that is, during acceptance testing. So the following diagram gives the DRE for the those values.
Now let's calculate system DRE of the above given project. In order to calculate the system DRE we need to take the number of defects found during the system divided by the defects found during system testing plus the defects found during acceptance testing. The following figure shows the system DRE calculation step by step.
Unit testing DRE calculation is similar to system testing DRE. As you can see from the following figure it's nothing but the number of defects found during unit testing divided by the number of defects found during unit testing plus the number of defects found during system testing.
One of the important factors to be noted while calculating unit testing DRE is that we need to exclude those defects which cannot be produced due to the limitations of unit testing. For instance, passing of data between components to each other. In unit testing, because we test it as a single unit, we can never reproduce defects which involve interaction. So such kinds of defects should be removed to get accurate test results. Test effectiveness is the measure of the bug-finding ability of our tests. In short, it measures how good the tests were. So effectiveness is the ratio of the measure of bugs found during testing to the total bugs found. Total bugs are the sum of new defects found by the user plus the bugs found in the test. The following figure explains the calculations in a pictorial format.
Defect age is also called a phase age or phage. One of the most important things to remember in testing is that the later we find a defect the more it costs to fix it. Defect age and defect spoilage metrics work with the same fundamental, i.e., how late you found the defect. So the first thing we need to define is what is the scale of the defect age according to phases. For instance, the following table defines the scale according to phases. So, for instance, requirement defects, if found in the design phase, have a scale of 1, and the same defect, if propagated until the production phase, goes up to a scale of 4.
Once the scale is decided now we can find the defect spoilage. Defect spoilage is defects from the previous phase multiplied by the scale. For instance, in the following figure we have found 8 defects in the design phase from which 4 defects are propagated from the requirement phase. So we multiply the 4 defects with the scale defined in the previous table, so we get the value of 4. In the same fashion we
calculate for all the phases. The following is the spoilage formula. It's the ratio of the sum of defects passed from the previous phase multiplied by the discovered phase then finally divided by the total number of defects. For instance, the first row shows that total defects are 27 and the sum of passed on defects multiplied by their factor is 8 (4 1 = 4 + 2 2 = 4). In this way we calculate for all phases and finally the total. The optimal value is 1. A lower value of spoilage indicates a more effective defect discovery process.
Unstable software: If the software is still under development and undergoing many changes automation testing will not be that effective. Once in a blue moon test scripts: Do not automate test scripts which will be run once in a while. Code and document review: Do not try to automate code and document reviews; they will just cause trouble.
Figure 126: What should not be automated All repetitive tasks which are frequently used should be automated. For instance, regression tests are prime candidates for automation because they're typically executed many times. Smoke, load, and performance tests are other examples of repetitive tasks that are suitable for automation. White box testing can also be automated using various unit testing tools. Code coverage can also be a good candidate for automation. The following figure shows, in general, the type of tests which can be automated.
(I) Which Automation Tools Have You Worked with and Can You Explain Them Briefly?
Note For this book we are using AutomatedQA as the tool for testing automation. So we will answer this question from the point of view of the AutomatedQA tool. You can install the AutomationQA tool and practice for yourself to see how it really works.
Figure 127: Candidates for automation In this answer we will be testing a tool called "WindowsFileSearch." This tool offers the following functionality:
This tool basically searches files by name and internal content of the file. It also has a wildcard search as well as an extension search which means we can search the file by extension, for instance, *.doc, *.exe, etc. To answer this answer in detail we have used the FileSearch application. You can
Note
experiment with any other application installed on your system such as a messenger or office application. Let's go step by step to learn to use the AutomatedQA tool to automate our testing process. First, start the tool by clicking all programs AutomatedQA TestComplete 5. Once the tool is started you will get a screen as shown here. We first need to create a new project by using the New Project menu as shown in the following figure.
Figure 128: Create a new project After clicking on the new project we will be prompted for what kind of testing we are looking at, i.e., load testing, general testing, etc. Currently, we will select only GeneralPurpose Test project. At this moment, you can also specify the project name, location, and the language for scripting (Select VBscript, currently).
Figure 129: Select the type of project Once the project name and path are given you will then be prompted with a screen as shown here. These are project items which we need to be included in your project depending on the testing type. Because currently we are doing a Windows application test we need to select the project items as shown in the figure. Please note events have to be selected compulsorily.
Figure 130: Select project items Once you have clicked finished you will get the Test Complete Project Explorer as shown here. The Test Complete Project Explorer is divided into three main parts: Events, Scripts, and TestedApps. Script is where all the programming logic is present. In TestedApps we add the applications that we want to test. So let's first add the application in TestedApps.
Figure 131: Project explorer In order to add the application EXE in TestedApps we need to right click on the TestedApps folder and click on New Item.
Figure 132: Add new applications to the project You will then be prompted with a screen as shown here. Browse to your application EXE file and add it to the TestedApps folder.
Figure 133: Add the EXE to your TestedApps folder Currently, we have added the WindowsFileSearch application. Now that your application is added we need to start recording our test. In order to start recording click on the button shown in the figure or push SHIFT + F1.
Once the recording toolbar is seen right click on the application added and run your test. In this scenario you can see the WindowsFileSearch application running. In this we have recorded a complete test in which we gave the folder name and keyword, and then tried to see if we were getting proper results. Your application being tested can be something different so your steps may vary.
Figure 135: Recording Once the test is complete you can stop the recording using the button on the recording toolbar. Once you stop, the recording tool will generate script of all your actions done as shown in the figure. You can view the programming script as shown here.
Figure 136: Script generated for the recording Once the script is recorded you can run the script by right clicking and running it. Once you run it the script tool will playback all the test steps which you recorded.
Figure 137: Running the recorded test If everything goes right you can see the test log as shown here which signifies that your script has run successfully.
Figure 139: Concept of load testing So once the tool captures the request and response, we just need to multiply the request and response with the virtual user. Virtual users are logical users which actually simulate the actual physical user by sending in the same request and response. If you want to do load testing with 10,000 users on an application it's practically impossible. But by using the load testing tool you only need to create 1000 virtual users.
(A) Can You Give An Example Showing Load Testing for Websites?
Note As said previously we will be using the AutomatedQA tool for automation in this book. So let's try to answer this question from the same perspective. You can get the tool from the CD provided with the book. The first step is to open a new project using TestComplete.
Figure 141: Create a new project After that select HTTP Load Testing from the project types.
Figure 142: Select HTTP load testing Once you click "OK" you will get different project items which you need for the project. For load testing only select three, i.e., Events, HTTP Load Testing, and Script as shown here.
Figure 143: Select the three items in load testing This project has the following items: Stations, Tasks, Tests, and Scripts. Stations basically define how many users the load testing will be performed for. Task has the request and response captured. Tests and Scripts have the Script which is generated when we record the automated test.
Figure 145: Project items You need to specify the number of virtual users, tasks, and the browser type such as Internet Explorer, Opera, etc.
Figure 146: Assign the number of virtual users and the browser As said previously the basic idea in load testing is the request and response which need to be recorded. That can be done by using the recording taskbar and clicking the icon shown.
Figure 147: Record HTTP task Once you click on the icon you need to enter the task name for it.
Figure 148: Specify task name In order to record the request and response the tool changes the proxy setting of the browser. So you can see from the screen here just click yes and let the next screen change the proxy settings.
Figure 150: Changing proxy settings Once the setting is changed you can then start your browser and make some requests and responses. Once that is done click on the stop button to stop the recording.
Figure 151: Stop the task once done The tool actually generates a script for the task recorded. You can see the script and the code generated in the following figure. To view the code you can double click on the Test2 script (here we have named it Test2 script).
Figure 152: Test2 created If you double click the test you can see the code.
Figure 153: Code generated for test Right click on the task and run it and you will see a summary report as shown in the figure.
(I) Can You Explain Table-Driven Testing? OR (I) How Can You Perform Data-Driven Testing Using Automated QA?
[This question is left to the user. Please install the tool and try for yourself.]
interview question book we limit ourselves to TPA which is the most preferred estimation methodology for black box testing. There are five methodologies most frequently used:
Top down according to budget WBS (Work Breakdown Structure) Guess and gut feeling Early project data TPA (Test Point Analysis)
Used to estimate only black box testing. Require function points as inputs.
Figure 154: Inputs for TPA come from function points Note In the following section we will look into how to estimate function points.
It's rare that someone will ask you to give the full definition of function points. They will rather ask about specific sections such as GSC, ILF, etc. The main interest of the interviewer will be how you use the function point value in TPA analysis. Function point analysis is mainly done by the development team so from a testing perspective you only need to get the function point value and then use TPA to get the black box
testing estimates.
Not e
This document contains material which has been extracted from the IFPUG Counting Practices Manual. It is reproduced in this document with the permission of IFPUG.
Function Point Analysis was developed first by Allan J. Albrecht in the mid-1970s. It was an attempt to overcome difficulties associated with lines of code as a measure of software size, and to assist in developing a mechanism to predict efforts associated with software development. The method was first published in 1979, then later in 1983. In 1984 Albrecht refined the method and since 1986, when the International Function Point User Group (IFPUG) was set up, several versions of the Function Point Counting Practices Manual have come out.
Not e
The best way to understand any complicated system is to break the system down into smaller subsystems and try to understand those smaller sub-systems first. In a function point you break complicated systems into smaller systems and estimate those smaller pieces, then total up all the subsystem estimates to come up with a final estimate.
The Following are some terms used in FPA: [Function Point analysis].
The first step in FPA is to define the boundary. There are two types of major boundaries: Internal Application Boundary External Application Boundary
We will give the features of external application boundaries and the internal application boundaries will be obvious. The external application boundary can be identified using the following litmus test: Does it have or will it have any other interface to maintain its data, which was not developed by you?. Example: Your Company is developing an "Accounts Application" and at the end of the accounting year, you have to report to the tax department. The tax department has its own website where companies can connect and report their tax transactions. Tax department applications have other maintenance and reporting screens developed by the tax software department. These maintenance screens are used internally by the tax department. So the tax online interface has other interfaces to maintain its data which is not within your scope, thus we can identify the tax website reporting as an external application.
Does your program have to go through a third party API or layer? In order for your application to interact with the tax department application your code has to interact with the tax department API. The best litmus test is to ask yourself if you have full access to the system. If you have full rights to make changes then it is an internal application boundary, otherwise it is an external application boundary.
(B) Can You Explain the Elementary Process? OR (B) Can You Explain the Concept of the Static and Dynamic Elementary Processes?
The Elementary Processes
As said previously FPA is about breaking huge systems into smaller pieces and analyzing them. Software applications are a combination of elementary processes.
Not e
An EP is the smallest unit of activity that is meaningful to a user. An EP must be selfcontained and leave the application in a consistent state.
An elementary process is not necessarily completely independent. So, we can define elementary process as small units of self-contained functionality from a user's perspective.
There are two types of elementary processes: Dynamic elementary Process. Static elementary Process.
The dynamic elementary process moves data from an internal application boundary to an external application boundary or vice-versa. Examples of dynamic elementary processes include: Input data screen where a user inputs data into the application. Data moves from the input screen inside the application. Transactions exported in export files in XML or any other standard. Display reports which can come from an external application boundary and an internal application boundary.
Examples of static elementary processes include: Static elementary process which maintains the data of the application either inside the application boundary or in the external application boundary.
For instance, in a customer maintenance screen maintaining customer data is a static elementary process. (I) Can You Explain FTR, ILF, EIF, EI, EO, EQ, and GSC?
Elements of Function Points
ILFs are logically related data from a user's point of view. They reside in the internal application boundary and are maintained through the elementary process of the application. ILFs can have a maintenance screen but not always. Do not make the mistake of mapping a one-to-one relationship between ILFs and the technical database design. This can make FPA go very wrong. The main difference between ILFs and a technical database is an ILF is a logical view and a database is a physical structure (technical design). Example: A supplier database design will have tables such as Supplier, Supplier Address, SupplierPhonenumbers, but from the ILF point of view you will only see the Supplier as logically they are all Supplier details.
These files are logically related data from the user point of view. EIFs reside in the external application boundary. EIFs are used only for reference purposes and are not maintained by internal applications. EIFs are maintained by external applications.
The following points are to be noted for RETs: An RET is a sub-group element data of ILF or EIF. If there is no sub-group of ILF then count the ILF itself as one RET. A group of RETs within ILF are logically related. Most likely with a parent-child relationship. Example: A supplier has multiple addresses and every address can have multiple phone numbers (see the following figure which shows a database diagram). So, Supplier, SupplierAddress, and SupplierPhoneNumber are a RETs.
Please note the whole database is one supplier ILF as all belong to one logical section. The RET quantifies the relationship complexity of ILF and EIF.
Data Element Types (DETs)
The following are the points to be noted for DETs counting: Each DET should be user recognizable. For example, in the previous figure we have kept the auto increment field (SupplierId) as the primary key. The SupplierId field
from a user point of view never exists at all, it's only from a software designing aspect, so does not qualify as a DET. DETs should be a non-recursive fields in ILF. DETs should not repeat in the same ILF again, and should be counted only once. Count foreign keys as one DET. SupplierId does not qualify as a DET but its relationship in the SupplierAddress table is counted as a DET. So Supplierid_fk in the SupplierAddress table is counted as a DET. The same holds true for "Supplieraddressid_fk".
The following points are to be noted for FTRs: An FTR is a file or data referenced by a transaction. An FTR should be an ILF or EIF. So count each ILF or EIF read during the process. If the EP is maintained as an ILF then count that as an FTR. So by default you will always have one FTR in any EP.
The following are points to be noted for EIs: EIs are dynamic elementary processes in which data is received from the external application boundary. Example: User interaction screens, when data comes from the User Interface to the Internal Application. EIs may maintain the ILF of the application, but it's not a compulsory rule. Example: A calculator application does not maintain any data, but still the screen of the calculator will be counted as EI. Most of the time user screens will be EI, but again it's not a hard and fast rule. Example: An import batch process running from the command line does not have a screen, but still should be counted as EI as it helps pass data from the external application boundary to the internal application boundary.
The following are points to be noted for EQs: An EQ is a dynamic elementary process in which result data is retrieved from one or more ILF or EIF. In this EP some input requests have to enter the application boundary. Output results exits the application boundary. EQ does not contain any derived data. Derived data means any complex calculated data. Derived data is not just mere retrieval data but are combined with additional
formula to generate results. Derived data is not part of ILF or EIF, they are generated on the fly.
Not e
EQ does not update any ILF or EIF. EQ activity should be meaningful from a user perspective. EP is self-contained and leaves the business in a consistent state. DET and processing logic is different from other EQs. Simple reports form good a base for EQs. There are no hard and fast rules that only simple reports are EQs. Simple view functionality can also be counted as an EQ.
The Following are points to be noted for EOs: EOs are dynamic elementary processes in which derived data crosses from the internal application boundary to the external application boundary. EO can update an ILF or EIF. The Process should be the smallest unit of activity that is meaningful to the end user in business. EP is self-contained and leaves the business in a consistent state. DET is different from other EOs. So this ensures that we do not count EOs twice. They have derived data or formulae calculated data.
The major difference between EO and EQ is that data passes across the application boundary. Example: Exporting accounts transactions to some external file format such as XML or some other format, which later the external accounting software can import. The second important difference is that EQ has non-derived data and EO has derived data.
General System Characteristics (GSC) Section
This section is the most important section. All the previously discussed sections relate only to applications. But there are other things also to be considered while making software, such as are you going to make it an N-Tier application, what's the performance level the user is expecting, etc. These other factors are called GSCs. These are external factors which affect the software and the cost of the software. When you submit a function point to a client, he normally will skip everything and go to the GSC section first. The GSC gives us something called the VAFs (Value Added Factors).
There are 14 points associated with (VAFs) and the associated rating tables:
Data Communications
How many communication facilities are there to aid in the transfer or exchange of information with the application or system?
Table 5: Data communication Open table as spreadsheet
Rating 0 1 2 3
Description Application uses pure batch processing or a stand-alone PC. Application uses batch processing but has remote data entry or remote printing. Application uses batch processing but has remote data entry and remote printing. Application includes online data collection or TP (Teleprocessing) front-end to a batch process or query system. Application is more than a front-end, but supports only one type of TP communications protocol. Application is more than a front-end, and supports more than one type of TP communications protocol.
Rating Description 0 Application does not aid the transfer of data or processing functions between components of the system. Application prepares data for end-user processing on another component of the system such as PC spreadsheets or PC DBMS.
Rating Description 2 Data is prepared for transfer, then is transferred and processed on another component of the system (not for end-user processing). Distributed processing and data transfer are online and in one direction only. Distributed processing and data transfer are online and in both directions. Processing functions are dynamically performed on the most appropriate component of the system.
3 4 5
Performance
Ratin g 0 1 Description No special performance requirements were stated by the user. Performance and design requirements were stated and reviewed but no special actions were required. Response time or throughput is critical during peak hours.No special design for CPU utilization was required. Processing deadline is for the next business day. Response time or throughput is critical during all business hours. No special design for CPU utilization was required. Processing deadline requirements with interfacing systems are constraining. In addition, stated user performance requirements are stringent enough to require performance analysis tasks in the design phase. In addition, performance analysis tools were used in the design, development,
Ratin g Description and/or implementation phases to meet the stated user performance requirements.
Heavily Used Configuration
How heavily used is the current hardware platform where the application will be executed?
Table 8: Heavily used configuration Open table as spreadsheet
Rating Description 0 1 No explicit or implicit operational restrictions are included. Operational restrictions do exist, but are less restrictive than a typical application. No special effort is needed to meet the restrictions. Some security or timing considerations are included. Specific processor requirement for a specific piece of the application is included. Stated operation restrictions require special constraints on the application in the central processor or a dedicated processor. In addition, there are special constraints on the application in the distributed components of the system.
2 3 4
Transaction Rate
Ratin g 0 1 Description No peak transaction period is anticipated. Peak transaction period (e.g., monthly, quarterly, seasonally, annually) is anticipated. Weekly peak transaction period is anticipated. Daily peak transaction period is anticipated. High transaction rate(s) stated by the user in the application requirements or service-level agreements are high enough to require performance analysis tasks in the design phase. High transaction rate(s) stated by the user in the application requirements or service-level agreements are high enough to require performance analysis tasks and, in addition, require the use of performance analysis tools in the design, development, and/or installation phases.
2 3 4
Rating 0 1 2 3
Description All transactions are processed in batch mode. 1% to 7% of transactions are interactive data entry. 8% to 15% of transactions are interactive data entry. 16% to 23% of transactions are interactive data entry.
Rating 4 5
Description 24% to 30% of transactions are interactive data entry. More than 30% of transactions are interactive data entry.
End-User Efficiency
Was the application designed for end-user efficiency? There are seven end-user efficiency factors which govern how this point is rated.
Table 11: End-user efficiency factor Open table as spreadsheet
YES OR NO End-user Efficiency Factor. 1 Navigational aids (e.g., function keys, jumps, dynamically generated menus). Menus. Online help and documents. Automated cursor movement. Scrolling. Remote printing (via online transactions). Preassigned function keys. Batch jobs submitted from online transactions. Cursor selection of screen data. Heavy use of reverse video, highlighting, colors underlining, and other indicators.
2 3 4 5 6 7 8 9 10
YES OR NO End-user Efficiency Factor. 11 12 13 14 15 16 Hard copy user documentation of online transactions. Mouse interface. Pop-up windows. As few screens as possible to accomplish a business function. Bilingual support (supports two languages; count as four items). Multilingual support (supports more than two languages; count as six items).
Ratin g 0 1 2 3 Description None of the above. One to three of the above. Four to five of the above. Six or more of the above, but there are no specific user requirements related to efficiency. Six or more of the above and stated requirements for end-user efficiency are strong enough to require design tasks for human factors to be included (for example, minimize keystrokes, maximize defaults, use of templates). Six or more of the above and stated requirements for end-user efficiency are strong enough to require use of special tools and processes to demonstrate that the
Ratin g 0 1 Description None of the above. Online update of one to three control files is included. Volume of updating is slow and recovery is easy. Online update of four or more control files is included. Volume of updating is low and recovery is easy. Online update of major internal logical files is included. In addition, protection against data lost is essential and has been specially designed and programmed in the system. In addition, high volumes bring cost considerations into the recovery process. Highly automated recovery procedures with minimum operator intervention are included.
3 4
Complex Processing
YES OR NO 1
Complex Processing Factor Sensitive control (for e.g., special audit processing) and/or application specific security processing. Extensive logical processing. Extensive mathematical processing. Much exception processing results in incomplete transactions that must be processed again, for example, incomplete ATM transactions caused by TP interruption, missing data values, or failed edits. Complex processing to handle multiple input/output possibilities, for example, multimedia, or device independence.
2 3 4
Rating 0 1 2 3 4 5
Description None of the above. Any one of the above. Any two of the above. Any three of the above. Any four of the above. All five of the above
Reusability
Ratin g 0 1 2 3 4 Description No reusable code. Reusable code is used within the application. Less than 10% of the application considers more than one user's needs. Ten percent or more of the application considers more than one user's needs. The application was specifically packaged and/or documented to ease re-use, and the application is customized by the user at a source-code level. The application was specifically packaged and/or documented to ease re-use, and the application is customized for use by means of user parameter maintenance.
Installation Ease
Ratin g 0 Description No special considerations were stated by the user, and no special setup is required for installation. No special considerations were stated by the user but special setup is required for installation. Conversion and installation requirements were stated by the user and conversion
Ratin g Description and installation guides were provided and tested. The impact of conversion on the project is not considered to be important. 3 Conversion and installation requirements were stated by the user, and conversion and installation guides were provided and tested. The impact of conversion on the project is considered to be important. In addition to 2 above, automated conversion and installation tools were provided and tested. In addition to 3 above, automated conversion and installation tools were provided and tested.
Operational Ease
How effective and/or automated are start-up, back-up, and recovery procedures?
Table 18: Operational ease Open table as spreadsheet
Ratin g 0 Description No special operational considerations other than the normal back-up procedures were stated by the user. One, some, or all of the following items apply to the application. Select all that apply. Each item has a point value of one, except where noted otherwise. Effective start-up, back-up, and recovery processes were provided, but operator intervention is required. Effective start-up, back-up, and recovery processes were provided, and no operator intervention is required (count as two items). The application minimizes the need for tape mounts. The application minimizes the need for paper handling.
14
Ratin g 5 Description The application is designed for unattended operation. Unattended operation means no operator intervention is required to operate the system other than to start-up or shut-down the application. Automatic error recovery is a feature of the application.
Multiple Sites
Was the application specifically designed, developed, and supported to be installed at multiple sites for multiple organizations?
Table 19: Multiple sites Open table as spreadsheet
Descriptio n 0 Rating User requirements do not require consider the needs of more than one user/installation site. Needs of multiple sites were considered in the design, and the application is designed to operate only under identical hardware and software environments. Needs of multiple sites were considered in the design, and the application is designed to operate only under similar hardware and/or software environments. Needs of multiple sites were considered in the design, and the application is designed to operate under different hardware and/or software environments. Documentation and support plans are provided and tested to support the application at multiple sites and the application is as described by 1 or 2. Documentation and support plans are provided and tested to support the application at multiple sites and the application is as described by 3.
Facilitate Change
Was the application specifically designed, developed, and supported to facilitate change? The following characteristics can apply to the application:
Table 20: Facilitates change factors Open table as spreadsheet
YES OR NO 0 1
Facilitates factors. None of above. Flexible query and report facility is provided that can handle simple requests; for example, and/or logic applied to only one internal logical file (counts as one item). Flexible query and report facility is provided that can handle requests of average complexity, for example, and/or logic applied to more than one internal logical file (counts as two items). Flexible query and report facility is provided that can handle complex requests, for example, and/or logic combinations on one or more internal logical files (counts as three items). Business control data is kept in tables that are maintained by the user with online interactive processes, but changes take effect only on the next business day. Business control data is kept in tables that are maintained by the user with online interactive processes and the changes take effect immediately (counts as two items)
Rating 0 1 2
Description None of the above. Any one of the above. Any two of the above.
Rating
Description
3 4 5
Any three of the above. Any four of the above. All five of the above
All of the above GSCs are rated from 05.Then VAFs are calculated from the equation below:
Not e
VAF = 0.65 + (sum of all GSC factor)/100). GSC has not been widely accepted in the software industry. Many software companies use unadjusted function points rather than adjusted. ISO has also removed the GSC section from its books and only kept unadjusted function points as the base for measurement.
The following are the look-up tables which will be referred to during counting.
Table 22: EI rating table Open table as spreadsheet
EI Rating Table Data Elements FTR Less than 2 Equal to 2 Greater than 2 1 to 4 3 3 4 5 to 15 3 4 4 Greater than 15 4 6 6
This table says that in any EI (External Input), if your DET count (Data Element) and FTR (File Type Reference) exceed these limits, then this should be the FP (Function Point). For
example, if your DET exceeds >15 and the FTR is greater than 2, then the function point count is 6. The following tables also show the same things. These tables should be there before us when we are doing function point counting. The best way is to put these values in Excel with the formula so that you only have to put the quantity in the appropriate section and you get the final value.
Table 23: EO rating table Open table as spreadsheet
EO Rating Table Data Elements FTR Less than 2 2 or 3 Greater than 2 1 to 5 4 4 5 6 to 19 4 5 7 Greater than 19 5 7 7
EQ Rating Table Data Elements FTR Less than 2 2 or 3 Greater than 2 1 to 5 3 3 4 6 to 19 3 4 6 Greater than 19 4 6 6
ILF Rating Table Data Elements RET 1 RET 2 to 5 Greater than 6 1 to 19 7 7 10 20 to 50 7 10 15 51 or more 10 15 15
This section will discuss the practical way of counting FPs to end up with the number of men/days on a project. 1. 2. 3. 4. 5. Count the ILF, EIF, EI, EQ, RET, DET, FTR (this is basically all sections discussed above): This whole FP count will be called the "unadjusted function point." Then put rating values 0 to 5 for all 14 GSCs. Add the total of all 14 GSCs to set the total VAF. The formula for VAF = 0.65 + (sum of all GSC factors/100). Finally, make the calculation of adjusted function points. Formula: Total function point = VAF * unadjusted function point. Make an estimation of how many function points you will cover per day. This is also called the "performance factor." On the basis of the performance factor, you can calculate men/days. Let's try to implement these details in a sample customer project.
We will be evaluating the customer GUI, so we will assume what the customer GUI is all about. The following is the scope of the customer screen: The customer screen will be as shown here. After inputting the customer code and customer name, they will be verified with a credit card check. The credit card check is an external system. Every customer can have multiple addresses. The customer will have add and update functionality.
There is one ILF in the customer screen: The customer ILF. There is one EIF in the form. Credit card system.
The following are the ILF counting rules: ILF are logically related data from the user's point of view. Customers and customer addresses belong logically to the customer category. ILF reside in the internal application boundary and are maintained through the elementary process of the application. Customer resides inside the application boundary as we have full access over it.
Description There are total 9 DETs, all add and update buttons, even the credit check button, the address list box, check box active, all text boxes. There is only one RET, the customer addresses. So according to the ILF ranking table
Total function
EI Credit Card Information Number of Number of DETs RETs Description The credit card information referenced is an EIF. Note this file is only referenced for the credit card check. There's only one textbox credit card number and hence one DET is put in the side column. And RET is 0. Looking at the rating table the total FP is 5. So according to the above ranking table Total function 5 1 1
It is a dynamic elementary process in which data is received from the external application boundary. Customer details are received from the external boundary, that is, the customer input screen. EI may maintain the ILF of the application, but it's not a compulsory rule. In this sample project the customer ILF is maintained. So there are two EI: one for add and one from update.
While counting EI I have seen many people multiply it by 3. That means we are going to do all CRUD functionality (ADD, UPDATE, and DELETE). Here the customer screen has add and update. We can say that 2 * 6; that = 12 FP for this EI customer. But later when someone refers to your FP sheet he will be completely lost.
Table 28: EI for the add customer Open table as spreadsheet
EI Add Customer Description There are total 9 DETs, all add and update buttons, even the credit check button, the address list box, check box active, all text boxes. There are 3 FTRs, one is the address and the second is the credit card information, and the third is the customer itself. So according to the above ranking table
Number of DETs
Number of FTRs
Total function
Description There are 9 total DETs, all add and update uttons, even the credit check button, the address list box, check box active, all text boxes. There are 3 FTRs, one is the address, the second is the credit card
Description information, and the third is the customer itself. So according to the above ranking table
Total function
The following are rules to recognize EO: Data should cross application boundaries and should involve complex logic.
Credit card check processes can be complex as credit card API complexity is still not known. Credit card information crosses from the credit card system to the customer system.
Table 30: EO to check the credit card Open table as spreadsheet
Description One DET credit card number and one RET credit card itself. Note if there are no RET we use a default of one. Look for RET counting rules defined in previous sections. So according to the above ranking table
Total function
The following are rules used to recognize EQ: EQ is a dynamic elementary process in which result data is retrieved from one or more ILFs or EIFs. For editing, the customer we will need to retrieve the customer details. In this EP some input requests have to enter the application boundary. The customer code is inputted from the same screen.
Output results exit the application boundary. The customer details are displayed while the customer is editing the customer data. EQ does not contain any derived data. The customer data which is displayed does not contain any complex calculations.
Table 31: EQ to display customer edit Open table as spreadsheet
Description There are 5 DETs to be retrieved: customer code, customer name, credit card number, active, customer address. Only customer details and customer address will be referenced. So according to the above ranking table
Total function
So now let's add the total FPs from the previous tables:
Table 32: Total of all function points. Open table as spreadsheet
Function Point ILF customer EO credit card check system EIF credit card information EI customer (add and update) EQ display customer edit information Total unadjusted function points
4 5 12 3 31
So the unadjusted FPs come to 31. Please note we refer to unadjusted function points as we have not accounted for other variance factors of the project (programmers leaving the job, languages, we, architecture, etc.). In order to make it the adjusted function points, we have to calculate and tabulate the GSCs and end up with the VAFs.
Table 33: GSC Open table as spreadsheet
GSCs Data communications Distributed data processing Performance Heavily used configuration Transaction rate Online data entry End-user efficiency Online update Complex processing Reusability Installation ease Operational ease Multiple sites Facilitate change Total
Value (05) 1 1 4 0 1 0 4 0 0 3 4 4 0 0 22
This factor affects the whole FP so be very careful with this factor. So now, calculating the adjusted FPs = VAFs * total unadjusted. Now we know that the complete FPs for the customer GUI is 27 FPs. Now calculating the efficiency factor, we say that we will complete 3 FPs per day, that is, 9 working days. So, the whole customer GUI is of 9 working days (note: do not consider Saturday and Sunday as working days). FPs = 0.87 * 31 = 26.97 = rounded to 27 FPs.
Before reading this section please refer to the different SDLC cycles in previous chapters. Quotations are heavily affected by which software lifecycle you follow because deliverables change according to the SLDC model the project manager chooses for the project. Example: For the waterfall model we will have requirement documents, design documents, source code, and testing plans. But for prototyping models, in addition to the documents above, we will also need to deliver the rough prototype. For the build and fix model we will not deliver any of the documents and the only document delivered will be the source code. So according to the SDLC model deliverables change and hence the quotation. We will divide the estimation across requirement, design, implementation (coding), and testing. How the estimation divides across all deliverables is all up to the project manager and his plans.
Table 34: Phase-wise distribution of effort Open table as spreadsheet
Percentage distribution effort 10% of total effort 20% of total effort 100% of total effort 10% of total effort
The above sample is 100% of the distribution of effort across various phases. But note that function points or any other estimation methodology only gives you the total execution estimation. So you can see in the above distribution we have given coding 100%. But as previously said it is up to the project manager to change according to scenarios. From the above function point estimation the estimation is 7 days. Let's try to divide it across all phases.
Table 35: Phase-wise effort distribution of man days Open table as spreadsheet
Percentage distribution effort Distribution of men/days acrossphases 10% of total effort 20% of total effort 60% of total effort 10% of total effort 0.9 days 1.8 days 7 days 0.9 days 10.6 days
The table shows the division of project men/days across the project. Now let's put down the final quotation. But first just a small comment about test cases. The total number of Test Cases = (Function Points) raised to a power of 1.2.
(A) How Can You Estimate the Number of Acceptance Test Cases in a Project?
The number of acceptance test cases = 1.2 * Function Points. 2025% of the total effort can be allocated to the testing phase. Test cases are nondeterministic. That means if the test passes it takes "X" amount of time and if it does not then to amend it takes "Y" amount of time.
Final Quotation
One programmer will work on the project for $1,000 a month. So his 10.6 days of salary comes to around $318.00. The following quotation format is a simple format. Every company has its own quotation format. http://www.microsoft.com/mac/resources/templates.aspx? pid=templates has a good collection of decent templates.
Table 36: Final bill Open table as spreadsheet
XYZ SOFTWARE COMPANY TNC Limited, Western road 17, California. Quotation number: 90 Date: 1/1/2004 Customer ID: Z- 20090DATAENTRY Quantity 1 Description Customer Project Discount 0% Taxable 0% Total $318.00
Quotation Valid for 100 days Goods delivery date within 25 days of half payment Quotation prepared by: XYZ estimation department Approved by: SPEG department XYZ CustomerSampleFP.xls is provided on the CD which has all the estimation details you need.
GSC Acceptance in Software Industry
GSC factors have been always a controversial topic. Most software companies do not use GSC, rather they baseline UAFP or construct their own table depending on company project history. ISO has also adopted function points as units of measurement, but they also use UAFP rather than AFP. Let's do a small experiment to view the relationships between FP, AFP, GSC, and VAF. In this experiment we will assume UAFP = 120 and then lot graph with GSC in increments of five. So the formula is VAF = 0.65 + (GS/100). Here's the table with the values in the formula plotted.
FP 78 84 90 96 102 108 114 120 126 132 138 144 150 156 162
GSC 0 5 10 15 20 25 30 35 40 45 50 55 60 65 70
The following are the observations from the table and plot: The graph is linear. It also captures that the nature of complexity is linear. If the GSC value is zero then the VAF is 0.65. So the graph starts from UAFP * 0.65.GSC = 35 AFP = UAFP. So the VAF = 1. When GSC < 35 then AFP > UAFP. That means complexity decreases. When GSC > 35 then AFP > UAFP. That means complexity increases.
Readers must be wondering why 0.65? There are 14 GSC factors from zero to five. So the maximum value of VAF = 0.65 + (70/100) = 1.35. So that VAF does not have any affect, i.e., UAFP = FP, the VAF should be one. The VAF will be one when the GSC is 35, i.e., half of 70. So, in order to complete value "1", value "0.65" is taken. Note that the value is 0.35 when the GSC is 35, to complete the one factor, "0.65" is required. But the following is the main problem related to the GSCs. The GSCs are applied throughout FPs even when some GSCs do not apply to whole function points. Here's the example to demonstrate the GSC problem. Let's take the 11th GSC factor "installation ease." The project is of 100 UAFP and there is no consideration of the installation done previously by the client so the 11th factor is zero.
Table 38: GSC with installation ease zero Open table as spreadsheet
GSC with installation easewith ZERO GSC Data communications Value (05) 1
Table 38: GSC with installation ease zero Open table as spreadsheet
GSC with installation easewith ZERO GSC Distributed data processing Performance Heavily used configuration Transaction rate Online data entry End-user efficiency Online update Complex processing Reusability Installation ease Operational ease Multiple sites Facilitate change Total Value (05) 1 4 0 1 0 4 0 0 3 0 4 0 0 18
VAF = 0.65 + (18/100) = 0.83. So the FPs = 100 * 0.83 = 83 function points. But later the client demanded a full-blown installation for the project with auto updating when the new version is released. So we change the installation ease to 5.
GSC with installation easewith 5 GSC Data communications Distributed data processing Performance Heavily used configuration Transaction rate Online data entry End-user efficiency Online update Complex processing Reusability Installation ease Operational ease Multiple sites Facilitate change Total Value (05) 1 1 4 0 1 0 4 0 0 3 5 4 0 0 23
So VAFs = 0.65 + (23/100) = 0.88 so the FPs = 100 * 0.88 = 88. The difference is only 5 FPs which in no way is a proper effort estimate. You cannot make an auto update for a
software version in 5 function points. Just think about downloading the new version, deleting the old version, updating any databases, structure changes, etc. So that's the reason GSCs are not accepted in the software industry. It is best to baseline your UAFPs.
Enhancement Function Points
Major software projects fail not because of programmers or project managers, but due to changing needs of customers. Function point groups have come out with a methodology called "Enhancement Function Points."
The Formula is as Follows:
Formulae of EFP (Enhanced Function Points) = (ADD + CHGA) * VAFA + (DELFP) * VAFB ADD: This is where new function points added. This value is achieved by counting all new EPs (elementary processes) given in a change request. CHGA: Function points which are affected due to CR. This value is achieved by counting all DET, FTR, ILF, EI, EO, and EQ which are affected. Do not count elements that are not affected. VAFA: This is a VAF which occurs because of CR. The example previously given was the desktop application that was changed to a web application so the GSC factor is affected. DELFP: When CR is used for removing some functionality this value is counted. It's rare that the customer removes functionalities, but if they ever do the estimator has to take note of it by counting the deleted elementary processes. VAFB: Again removal affects value added factors.
Once we are through with calculating enhanced function points, it is time to count the total function points of the application. The formula is as follows: Total Function points = [UFPB + ADD + CHGA]-[CHGB-DELFP] UFPB: Function points previously counted before enhancement. ADD: Newlyadded functionality which leads to new function points after enhancements. CHGA: Changed function points counted after enhancements. CHGB: Changed function points before enhancements. DELFP: Deleted function points.
There are three main elements which determine estimates for black box testing: size, test strategy, and productivity. Using all three elements we can determine the estimate for black box testing for a given project. Let's take a look at these elements. Size: The most important aspect of estimating is definitely the size of the project. The size of a project is mainly defined by the number of function points. But a function point fails or pays the least attention to the following factors:
Complexity: Complexity defines how many conditions exist in function points identified during a project. More conditions means more test cases which means more testing estimates. For instance, the following is an application which takes the customer name from the end user. If the customer name is greater than 20 characters then the application should give an error. So for this case there will be one test case. But let's say the end user also puts one more condition that if the user inputs any invalid character then the application should give an error. Because there is one more extra condition in the project the complexity has increased, which also means that we need to test two cases. The following illustrates this figure. Interfacing: How much does one function affect the other part of the system? If a function is modified then accordingly the other systems have to be tested as one function always impacts another. Uniformity: How reusable is the application? It is important to consider how many similar structured functions exist in the system. It is important to consider the extent to which the system allows testing with slight modifications.
Figure 159: Complexity Test strategy: Every project has certain requirements. The importance of all these requirements also affects testing estimates. Any requirement importance is from two perspectives: one is the user importance and the other is the user usage. Depending on these two characteristics a requirement rating can be generated and a strategy can be chalked out accordingly, which also means that estimates vary accordingly. Productivity: This is one more important aspect to be considered while estimating black box testing. Productivity depends on many aspects. For instance, if your project has new testers your estimates shoot up because you will need to train the new testers in terms of project and domain knowledge. Productivity has two important aspects: environment and productivity figures. Environmental factors define how much the environment affects a
project estimate. Environmental factors include aspects such as tools, test environments, availability of testware, etc. While the productivity figures depend on knowledge, how many senior people are on the team, etc. The following diagram shows the different elements that constitute TPA analysis as discussed.
On the CD we have provided an Excel file called "FunctionPoints (Accounting Application)", which is used in this example to make TPA calculation easier. We have also provided an explanation of how to use the Excel spread sheet. The entire image snapshot which you will see in this answer is taken from the "FunctionPoints file."
In order to really answer this question let's do one complete estimation practically for a sample project. The following is a simple accounting application developed for http://www.questpond.com to track its sales. The first screen is a voucher entry screen. It's a normal simple voucher entry screen with extra functionality to print the voucher. The second screen is a master screen for adding accounting codes.
The following are point requirements gathered from the end customer: 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. 15. 16. 17. 18. The account code entered in the voucher entry screen should be a valid account code from the defined chart of accounts given by the customer. The user should be able to add, delete, and modify the account code from the chart of the account master (this is what the second screen defines). The user will not be able to delete the chart of account codes if he has already entered transactions for in vouchers. The chart of account code master will consist of account codes and descriptions of the account code. The account code cannot be greater than 10. The voucher data entry screen consists of the debit account code, credit account code, date of transaction, and amount. Once the user enters voucher data he should be able to print it in the future at any time. The debit and credit account are compulsory. The amount value should not be negative. After pressing the submit button the value should be seen in the grid. The amount is compulsory and should be more than zero. The debit and credit account should be equal in value. Only numeric and non-negative values are allowed in the amount field. Two types of entries are allowed: sales and commissions. Date, amount, and voucher number are compulsory. The voucher number should be in chronological order and the system should auto increment the voucher number with every voucher added. No entries are allowed for previous months. Users should be able to access data from separate geographical locations. For instance, if one user is working in India and the other in China, then both users should be able to access each other's data through their respective location.
Now that we have all the requirements let's try to estimate how we can use TPA to do get the actual men days needed to complete a project. The following figure shows our road map and how we will achieve using TPA. There are in all ten steps needed to achieve our goal.
Figure 163: TPA steps. Step 1: Calculate Function Points Not You will understand this section if you have not read the function points explanation e
given previously.
EI Calculation
The following are the EI entries for the accounting application. Currently, we have two screens: one is the master screen and one is the voucher transaction screen. In the description we have also described which DETs we have considered. For the add voucher screen we have 7 DETs (note the buttons are also counted as DETs) and for the account code master we have 4 DETs.
There are no EIFs in the system because we do not communicate with any external application.
EOs are nothing but complex reports. In our system we have three complex reports: trial balance, profit and loss, and balance sheet. By default we have assumed 20 fields which makes it a complex report (when we do estimations sometimes assumptions are okay).
EQs are nothing but simple output sent from the inside of the application to the external world. For instance, a simple report is a typical type of EQ. In our current accounting application we have one simple form that is the print voucher. We have assumed 20 DETs so that we can move ahead with the calculation.
As said in the FPA tutorial previously given the GSC factor defines the other factors of the projects which the FP counting does not accommodate. For the accounting application we have kept all the GSC factors as 1 except for data communications and performance. We have kept communication as 2 because, one the requirement point is that we need application data to be accessed from multiple centers which increases the data communication complexity and also because the requirement of the end customer is that performance should be mediumly good. The following figure shows the GSC entries.
Figure 168: GSC factors for our accounting application Total Calculation
Now that we have filled in all the details we need to calculate the total man days. The following figure explains how the calculations are done. The first five rows, i.e., ILF, EIF, EO, EQ, and EI, are nothing but the total of the individual entries. A total unadjusted function point is the total of ILF + EIF + EO + EQ + EI. We get the total adjusted function which is nothing but the total un-adjusted function points multiplied by the GSC factors. Depending on the organizations baseline we define how many FPs can be completed by a programmer in one day. For instance, for the following accounting application we have 1.2 FPs per day. Depending on the FPs per day we get the total man days. Once we have the total man days we distribute these values across the phases. We have just found the total execution time. So we have assigned the total man days to the execution phase. From the execution phase and
man days we distribute 20 percent to the requirement phase, 20 percent to technical design, and 5 percent to testing.
Now that we have completed the function point analysis for this project let's move on to the second step needed to calculate black box testing using TPA.
Step 2: Calculate Df (Function-Dependant Factors)
Df is defined for each function point. So all the function point descriptions as well as values are taken and the Dfs are calculated for, each of them. You can see from the figure how every function point factor is taken, and how the Df is calculated.
But we have still not seen how Df will be calculated. Df is calculated using four inputs: user importance, usage intensity, interfacing, and complexity. The following figure shows the different inputs in a pictorial manner. All four factors are rated as low, normal, and high and assigned to each function are factors derived from the function points. Let's take a look at these factors.
User importance (Ue): How important is this function factor to the user compared to other function factors? The following figure shows how they are rated. Voucher data, print voucher, and add voucher are rated with high user importance. Without these the user cannot work at all. Reports have been rated low because they do not really stop the user from working. The chart of accounts master is rated low because the master data is something which is added at one time and can also be added from the back end.
Usage intensity (Uy): This factor tells how many users use the application and how often. The following figure shows how we have assigned the values to each function factor. Add voucher, Print Voucher, and voucher data are the most used function factors. So they are rated high. All other function factors are rated as low.
Interfacing (I): This factor defines how much impact this function factor has on other parts of the system. But how do we now find the impact? In TPA, the concept of LDS is used to determine the interfacing rating. LDS stands for Logical Data Source. In our project we have two logical data sources: one is voucher data and the other is account code data (i.e., chart of accounts data). The following are the important points to be noted which determine the interfacing: We need to consider only functions which modify LDS. If a function is not modifying LDS then its rating is Low by default. To define LDS we need to define how many LDSs are affected by the function and how many other functions access the LDS. Other functions only need to access the function; even if they do not modify it.
The following is the table which defines the complexity level according to the number of LDSs and functions impacting on LDS.
So now depending on the two points defined above let's try to find out the interfacing value for our accounting project. As said previously we have two functions which modify LDS in our project: one is the add voucher function which affects the voucher data and the add account code which affects the chart of accounts code (i.e., the accounts code master). The add voucher function primarily affects voucher data LDFs. But other functions such as reports and print also use the LDS. So in total there are five functions and one LDS. Now looking at the number of LDSs and the number of functions the impact complexity factor is Low.
The other function which modifies is the Add account code function. The LDS affected is the chart of account code and the function which affects it is the Add account code function. There are other functions that indirectly affect this function, too. For instance, Report which needs to access account code, Print voucher which uses the account code to print account description and also the Add voucher function which uses the chart of accounts code LDS to verify if the account code is correct. So we can see the look-up table and the impact complexity factor is Average.
The other function factors do not modify any data so we give them a Low rating. The following is the interfacing complexity factors assigned.
Complexity (C): This factor defines how complex the algorithm for the particular function factor is. Add voucher is the most complex function in the project and it can have more than 11 conditions so we have rated the complexity factor the highest. Reports are mildly complex and can be rated as average in complexity. So as discussed we have assigned values accordingly as shown in the figure.
Uniformity (U): This factor defines how reusable a system is. For instance, if a test case written for one function can be applied again then it affects the testing estimates accordingly. Currently, for this project, we have taken a uniformity factor of 1. So, for example, if the customer had a requirement to also update the accounts code then we could have used two functions, Add voucher and Update voucher.
One we have all the five factors we apply the following formula to calculate Df for all the function factors: Df = [(Ue + Uy + I + C)/16] * U
Step 3: Calculate Qd
The third step is to calculate Qd. Qd, i.e, dynamic quality characteristics, have two parts: explicit characteristics (Qde) and implicit characteristics (Qdi). Qde has five important characteristics: Functionality, Security, Suitability, Performance, and Portability. The following diagram shows how we rate those ratings. Qdi defines the implicit characteristic part of the Qd. These are not standard and vary from project to project. For instance, we have identified for this accounting application four characteristics: user friendly, efficiency, performance, and maintainability. From these four characteristics we assign each 0.02 value. We can see from the following figure for user friendliness we have assigned 0.02 value. In the Qde part we have given functionality normal importance and performance as relatively unimportant but we do need to account for them. Once we have Qde and Qdi then Qd = Qde
+ Qdi. For this sample you can see that the total value of Qd is 1.17 (which is obtained from 1.15 + 0.02). Qd is calculated using the rating multiplied by the value. The following table shows the rating and the actual value. So the 1.15 has come from the following formula: ((5 * 0.75) + (3 * 0.05) + (4 * 0.10) + (3 * 0.10) + (3 * .10)) / 4
Figure 182: Calculation of Qd (dynamic characteristics) Step 4: Calculate TPf for Each Function
In this step we calculate TPf (number of test points assigned to each function). This is done by using three data values (FPf, Df, and Qd). The following is the formula: TPf = FPf * Df * Qd
Because we are using the Excel worksheet these calculations are done automatically. The following figure shows how the TPf calculations are done.
In this step we take into account the static quality characteristic of the project. This is done by defining a checklist of properties and then assigning a value of 16 to those properties. For this project we have only considered easy-to-use as a criteria and hence assigned 16 to it.
Now that we have TPfs for all function factors, FPs and Qs (static test point data), it's time to calculate the Tp (Total number of test points). The formula is as follows: Tp = sum(TPf) + (FPQs/500)
For the accounting system total Tp = 71.21 (use a calculator if needed). The following figure shows how the total Tp is derived.
Productivity/skill factors show the number of test hours needed per test points. It's a measure of experience, knowledge, and expertise and a team's ability to perform. Productivity factors vary from project to project and also organization to organization. For instance, if we have a project team with many seniors then productivity increases. But if we have a new testing team productivity decreases. The higher the productivity factor the higher the number of test hours required. For this project we have good resources with great ability. So we have entered a value of 1.50 which means we have high productivity.
Figure 186: Productivity factor/Skill factor Step 8: Calculate Environmental Factor (E)
The number of test hours for each test point is influenced not only by skills but also by the environment in which those resources work. The following figure shows the different environmental factors. You can also see the table ratings for each environmental factor.
Figure 192: Development environment Step 9: Calculate Primary Test Hours (PT)
Primary test hours are the product of test points, skill factors, and environmental factors. The following formula shows the concept in more detail: Primary test hours = TP * Skill factor * E
For the accounting application the total primary test hours is 101.73 as shown in the figure.
Figure 193: Primary test hours Step 10: Calculate Total Hours
Every process involves planning and management activities. We also need to take into account these activities. Planning and management is affected by two important concepts. Team size and management tools. So below are the rating sheet for team size and management tools. These values are summed and the percentage of this value is then multiplied with the primary test hours.
Finally, we distribute this number across the phases. So the total black box testing estimate for this project in man hours is 101.73 man hours, a 13-man day approximately.