0% found this document useful (0 votes)
150 views2 pages

Ca&o Unit 1 Day1 PDF

There are five generations of computers defined by changes in technology. The first generation used vacuum tubes, the second used transistors which were smaller and more efficient than tubes. The third generation saw the development of integrated circuits and interaction through keyboards and monitors. The fourth generation was defined by the invention of microprocessors, putting all components on a single chip. The fifth generation involves advances in artificial intelligence through technologies like parallel processing and superconductors.

Uploaded by

aadi1988
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
150 views2 pages

Ca&o Unit 1 Day1 PDF

There are five generations of computers defined by changes in technology. The first generation used vacuum tubes, the second used transistors which were smaller and more efficient than tubes. The third generation saw the development of integrated circuits and interaction through keyboards and monitors. The fourth generation was defined by the invention of microprocessors, putting all components on a single chip. The fifth generation involves advances in artificial intelligence through technologies like parallel processing and superconductors.

Uploaded by

aadi1988
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 2

Generations

Generation in computer terminology is a change in technology a computer is/was being


used. Initially, the generation term was used to distinguish between varying hardware
technologies. Nowadays, generation includes both hardware and software, which together
make up an entire computer system.

There are five computer generations known till date. Each generation has been discussed
in detail along with their time period and characteristics. In the following table,
approximate dates against each generation have been mentioned, which are normally
accepted.

1940 – 1956: First Generation – Vacuum Tubes

These early computers used vacuum tubes as circuitry and magnetic drums for memory.
As a result they were enormous, literally taking up entire rooms and costing a fortune to
run. These were inefficient materials which generated a lot of heat, sucked huge electricity
and subsequently generated a lot of heat which caused ongoing breakdowns.

These first generation computers relied on ‘machine language’ (which is the most basic
programming language that can be understood by computers). These computers were
limited to solving one problem at a time. Input was based on punched cards and paper
tape. Output came out on print-outs.

1956 – 1963: Second Generation – Transistors

The replacement of vacuum tubes by transistors saw the advent of the second generation
of computing. Although first invented in 1947, transistors weren’t used significantly in
computers until the end of the 1950s. They were a big improvement over the vacuum
tube, despite still subjecting computers to damaging levels of heat. However they were
hugely superior to the vacuum tubes, making computers smaller, faster, cheaper and less
heavy on electricity use. They still relied on punched card for input/printouts.

The language evolved from cryptic binary language to symbolic (‘assembly’) languages.
This meant programmers could create instructions in words. About the same time high
level programming languages were being developed (early versions of COBOL and
FORTRAN). Transistor-driven machines were the first computers to store instructions into
their memories – moving from magnetic drum to magnetic core ‘technology’. The early
versions of these machines were developed for the atomic energy industry.

1964 – 1971: Third Generation – Integrated Circuits

By this phase, transistors were now being miniaturised and put on silicon chips (called
semiconductors). This led to a massive increase in speed and efficiency of these
machines. These were the first computers where users interacted using keyboards and
monitors which interfaced with an operating system, a significant leap up from the punch
cards and printouts. This enabled these machines to run several applications at once
using a central program which functioned to monitor memory.

As a result of these advances which again made machines cheaper and smaller, a new
mass market of users emerged during the ‘60s.

1972 – 2010: Fourth Generation – Microprocessors

This revolution can be summed in one word: Intel. The chip-maker developed the Intel
4004 chip in 1971, which positioned all computer components (CPU, memory, input/output
controls) onto a single chip. What filled a room in the 1940s now fit in the palm of the
hand. The Intel chip housed thousands of integrated circuits. The year 1981 saw the first
ever computer (IBM) specifically designed for home use and 1984 saw the MacIntosh
introduced by Apple. Microprocessors even moved beyond the realm of computers and
into an increasing number of everyday products.

The increased power of these small computers meant they could be linked, creating
networks. Which ultimately led to the development, birth and rapid evolution of the
Internet. Other major advances during this period have been the Graphical user interface
(GUI), the mouse and more recently the astounding advances in lap-top capability and
hand-held devices.

2010- : Fifth Generation – Artificial Intelligence

Computer devices with artificial intelligence are still in development, but some of these
technologies are beginning to emerge and be used such as voice recognition.

AI is a reality made possible by using parallel processing and superconductors. Leaning to


the future, computers will be radically transformed again by quantum computation,
molecular and nano technology.

The essence of fifth generation will be using these technologies to ultimately create
machines which can process and respond to natural language, and have capability to
learn and organise themselves.

You might also like