Performance Testing For Best Case: International Journal of Pavement Engineering July 2021
Performance Testing For Best Case: International Journal of Pavement Engineering July 2021
net/publication/353482454
CITATIONS READS
0 816
1 author:
SEE PROFILE
All content following this page was uploaded by Jun Luke Cui on 27 July 2021.
Performance Testing is a software testing process used for testing the speed, response time, stability,
reliability, scalability, and resource usage of a software application under a particular workload. The main
purpose of performance testing is to identify and eliminate the performance bottlenecks in the software
application. It is a subset of performance engineering and is also known as “Perf Testing”.
Features and Functionality supported by a software system is not the only concern. A software
application's performance like its response time, reliability, resource usage and scalability do matter. The
goal of Performance Testing is not to find bugs but to eliminate performance bottlenecks.
Performance Testing is done to provide stakeholders with information about their application regarding
speed, stability, and scalability. More importantly, Performance Testing uncovers what needs to be
improved before the product goes to market. Without Performance Testing, the software is likely to suffer
from issues such as: running slow while several users use it simultaneously, inconsistencies across
different operating systems, and poor usability.
Performance testing will determine whether their software meets speed, scalability, and stability
requirements under expected workloads. Applications sent to the market with poor performance metrics
due to nonexistent or poor performance testing are likely to gain a bad reputation and fail to meet
expected sales goals.
• Load testing - checks the application's ability to perform under anticipated user loads. The
objective is to identify performance bottlenecks before the software application goes live.
• Stress testing - involves testing an application under extreme workloads to see how it handles
high traffic or data processing. The objective is to identify the breaking point of an application.
• Endurance testing - is done to make sure the software can handle the expected load over a long
period of time.
• Spike testing - tests the software's reaction to sudden large spikes in the load generated by users.
• Volume testing - Under Volume Testing large no. of. Data is populated in a database and the
overall software system's behavior is monitored. The objective is to check software application's
performance under varying database volumes.
• Scalability testing - The objective of scalability testing is to determine the software application's
effectiveness in "scaling up" to support an increase in user load. It helps plan capacity addition to
your software system.
Most performance problems revolve around speed, response time, load time, and poor scalability. Speed is
often one of the most important attributes of an application. A slow-running application will lose potential
users. Performance testing is done to make sure an app runs fast enough to keep a user's attention and
interest. Take a look at the following list of common performance problems and notice how speed is a
common factor in many of them:
• Long Load time - Load time is normally the initial time it takes an application to start. This should
generally be kept to a minimum. While some applications are impossible to make load in under a
minute, Load time should be kept under a few seconds if possible.
• Poor response time - Response time is the time it takes from when a user inputs data into the
application until the application outputs a response to that input. Generally, this should be very
quick. Again if a user has to wait too long, they lose interest.
• Poor scalability - A software product suffers from poor scalability when it cannot handle the
expected number of users or when it does not accommodate a wide enough range of users. Load
Testing should be done to be certain the application can handle the anticipated number of users.
• Bottlenecking - Bottlenecks are obstructions in a system that degrade overall system
performance. Bottlenecking is when either coding errors or hardware issues cause a decrease of
throughput under certain loads. Bottlenecking is often caused by one faulty section of code. The
key to fixing a bottlenecking issue is to find the section of code that is causing the slow down and
try to fix it there. Bottlenecking is generally fixed by either fixing poor running processes or adding
additional Hardware. Some common performance bottlenecks are
The methodology adopted for performance testing can vary widely but the objective for performance tests
remain the same. It can help demonstrate that your software system meets certain pre-defined
performance criteria. Or it can help compare the performance of two software systems. It can also help
identify parts of your software system which degrade its performance.
1. Identify your testing environment - Know your physical test environment, production environment
and what testing tools are available. Understand details of the hardware, software and network
configurations used during testing before you begin the testing process. It will help testers create
more efficient tests. It will also help identify possible challenges that testers may encounter during
the performance testing procedures.
2. Identify the performance acceptance criteria - This includes goals and constraints for throughput,
response times and resource allocation. It is also necessary to identify project success criteria
outside of these goals and constraints. Testers should be empowered to set performance criteria
and goals because often the project specifications will not include a wide enough variety of
performance benchmarks. Sometimes there may be none at all. When possible finding a similar
application to compare to is a good way to set performance goals.
3. Plan & design performance tests - Determine how usage is likely to vary amongst end users and
identify key scenarios to test for all possible use cases. It is necessary to simulate a variety of end
users, plan performance test data and outline what metrics will be gathered.
4. Configuring the test environment - Prepare the testing environment before execution. Also,
arrange tools and other resources.
5. Implement test design - Create the performance tests according to your test design.
6. Run the tests - Execute and monitor the tests.
7. Analyze, tune and retest - Consolidate, analyze and share test results. Then fine tune and test
again to see if there is an improvement or decrease in performance. Since improvements generally
• Verify response time is not more than 4 secs when 1000 users access the website simultaneously.
• Verify response time of the Application Under Load is within an acceptable range when the
network connectivity is slow
• Check the maximum number of users that the application can handle before it crashes.
• Check database execution time when 500 records are read/written simultaneously.
• Check CPU and memory usage of the application and the database server under peak load
conditions
• Verify response time of the application under low, normal, moderate and heavy load conditions.
During the actual performance test execution, vague terms like acceptable range, heavy load, etc. are
replaced by concrete numbers. Performance engineers set these numbers as per business requirements,
and the technical landscape of the application.