Computer System General requirements
Computer System General requirements
- Reliability:
- Fault tolerant
- Scalable
- Security
- Upgrade and Update
1. Reliability
1
ROCOF (rate of failure occurrence): The number of unexpected events
over a particular time of operation. A firewall that crashes an average of
five times every 1,000 hours has a ROCOF of 5 per 1,000 hours.
Mean Time To Restore (MTTR) :is the average time taken to reinstate
a failed component to functioning state. In previous example if the first
if the first time we took 1 hor and 2 hours in the rest times, so the
MTTR = 1.8 h
Where MTBF is Mean Time Before Failures and MTTR is Mean Time
To Repair.
2
MTBF for serial and parallel systems:
3
2. Fault tolerant
To improve your system reliability you should be ready when a failure occurs.
This failure could be:
Software failures
Hardware failures
Human errors
Disasters
The fault tolerant reflects the ability to continue non-stop when a failure occurs. A
fault-tolerant system is designed from the ground up for reliability by building
multiples of all critical components, such as CPUs, memories, disks and power
supplies into the same computer. In the event one component fails, another
takes over without skipping a beat.
True fault tolerant systems with redundant hardware are the most costly because
the additional components add to the overall system cost. However, fault tolerant
systems provide the same processing capacity after a failure as before, whereas
high availability systems often provide reduced capacity.
There are many level of fault tolerant could be expressed by the following terms:
hot swap : To pull out a component from a system and plug in a new one
while the main power is still on. Also called "hot plug" and "hot insertion.
4
in Fault Tolerant Environments Hot swap is a desired feature of fault
tolerant systems built with redundant drives, circuit boards and power
supplies, especially servers that run 24/7. When a component fails and the
redundant unit takes over, the bad one can be replaced without stopping
the operation.
3. Clustering
Using two or more computer systems that work together. It generally refers to
multiple servers that are linked together in order to handle variable workloads
or to provide continued operation in the event one fails. Each computer may
be a multiprocessor system itself. For example, a cluster of four computers,
each with four CPUs, would provide a total of 16 CPUs processing
simultaneously.
A cluster of servers provides fault tolerance and/or load balancing. If one server fails, one or more
additional servers are still available. Load balancing distributes the workload over multiple
systems.
4. Scalable
5
Expandable or scalable. Referring to hardware or software, the term has
become a popular buzzword in the IT world. A "highly scalable" device or
application implies that it can handle a large increase in users, workload or
transactions without undue strain.
Scalable does not always mean that expansion is free ( like adding user to a
domain for example). Extra-cost hardware or software may be required to
achieve maximum scalability. Nevertheless, scalability is a positive feature of
a product sold to fast-growing companies.
5. Security:
(Later in chapter 14 )
6. Upgrade and Update
Usually we refer to the upgrade by replacing an existing item by a newer one.
In general, we use this term with hardware and software. While we use
update to talk about replacing existing data or file by the most recent available
one. Some vendor talk about updating when we replace an item or few while
they use upgrade when the make a hall change.
The Servers:
1. Servers Classifications:
There is no general classification used to classify the servers, this
classification refers to the vendors. The main vendors in the market today are:
IBM, HP, SUN,DELL, Fujitsu and others.
In general, and to classify the servers, we could use the following criteria:
6
o HP : HP In today’s market, HP has many server brands
HP Integrity servers
HP Integrity NonStop servers
HP 9000 servers
HP AlphaServer systems
HP e3000 servers
Refer to www.hp.com for more information
Performance:
The speed with which a computer processes data. It is a combination of internal
processing speed, peripheral speeds (I/O) and the efficiency of the operating
system and other system software all working together.
Throughput
1) In computer technology, throughput is the amount of work that a computer can
do in a given time period. Historically, throughput has been a measure of the
comparative effectiveness of large commercial computers that run many
programs concurrently. An early throughput measure was the number of batch
jobs completed in a day. More recent measures assume a more complicated
mixture of work or focus on some particular aspect of computer operation. While
"cost per million instructions per second (MIPS)" provides a basis for comparing
the cost of raw computing over time or by manufacturer, throughput theoretically
tells you how much useful work the MIPS are producing.
7
amount of time between a single interactive user request being entered and
receiving the application's response is known as response time.
Response time
MIPS
The number of MIPS (million instructions per second) is a general measure of
computing performance and, by implication, the amount of work a larger
computer can do. For large servers or mainframes, MIPS is a way to measure
the cost of computing: the more MIPS delivered for the money, the better the
value. Historically, the cost of computing measured in the number of MIPS has
been reduced by half on an annual basis for a number of years.
The number of MIPS attributed to a computer is usually determined by one or
more benchmark runs
performance
Performance seems to have two meanings:
1) The speed at which a computer operates, either theoretically (for example,
using a formula for calculating Mtops - millions of theoretical instructions per
second) or by counting operations or instructions performed (for example, (MIPS)
- millions of instructions per second) during a benchmark test. The benchmark
test usually involves some combination of work that attempts to imitate the kinds
of work the computer does during actual use. Sometimes performance is
expressed for each of several different benchmarks.
8
2) The total effectiveness of a computer system, including throughput, individual
response time, and availability.