0% found this document useful (0 votes)
2 views13 pages

Week8e Notecomputersss3datarepresentation

The document discusses data representation, defining it as the method of representing data in a physical storage medium using binary digits (bits). It outlines four basic coding systems: Bits, BCD, ASCII, and EBCDIC, explaining how each represents data. Additionally, it covers computer character sets and terms related to data representation, such as typecasting and parity bits.

Uploaded by

jibakin4life
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
2 views13 pages

Week8e Notecomputersss3datarepresentation

The document discusses data representation, defining it as the method of representing data in a physical storage medium using binary digits (bits). It outlines four basic coding systems: Bits, BCD, ASCII, and EBCDIC, explaining how each represents data. Additionally, it covers computer character sets and terms related to data representation, such as typecasting and parity bits.

Uploaded by

jibakin4life
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 13

WEEK 8

DATA REPRESENTATION
CONTENTS:
1. Definition of data representation.
2. List methods of data representation:
(i) Bits
(ii) BCD
(i) EBCDIC
(ii) ASCII
3. Represent data in different character sets
(Computer Character sets)

WEEK 8
DATA REPRESENTATION
WHAT IS DATA REPRESENTATION?
Data Representation or Coding System is the way data is
represented in a physical storage medium.

Data representation is the way that the physical properties of a


medium are used to represent data and the manner in which data
is expressed symbolically by digits in a computer.

Data representation refers to the methods used internally to


represent information stored in a computer.

Data representation also refers to the manner in which data is


expressed symbolically by binary digits in a computer.

Computers store lots of different types of information such as


numbers, texts, and graphics of many varieties (still, videos,
audios, sound, and animations) e.t.c.

In Computer, data are represented in Binary Digits called Bits, that


is, 0s and 1s. Bits are usually referred to as a single element in a
computer. They are combined in groups of eight called Byte.
Computer “word” information usually 8-bit, 16-bit, 32-bit, 64-bit to
be handled by a computer are coded into 1s and 0s.

In data representation, the following equivalents are used:

8 bits = 1 byte
210 bytes = 1024 bytes = 1 Kilobyte (KB)
220 bytes = 1024 KB = 1 Megabyte (MB)
230 bytes = 1024 MB = 1 Gigabyte (GB)
240 bytes = 1024 GB = 1 Terabyte (TB)
250 bytes = 1024 TB = 1 Petabyte (PB)

METHODS OF DATA
REPRESENTATION/CODING
SYSTEM
There are four (4) basic coding systems or methods of data
representation:
1. Bits (Binary Digits)

2. BCD (Binary Coded Decimal)

3. ASCII (America Standard Code for Information Interchange)

4. EBCDIC (Extended Binary Coded Decimal Interchange Code)

RESEARCH AND ASSIGNMENT

Explain the four (4) basic coding systems or methods of data


representation:

EXPLANATION AND
SOLUTION
1. Bits (Binary Digits)
In computer system, all types of information are stored
internally in the same simple format, that is, a sequence of 0s
and 1s. This is referred to as BITS.

A bit means Binary digit which is represented by “0” and “1”.

A bit is the basic unit of information in computing and


telecommunications. It is the amount of information stored by
a digital device or other physical system that exists in one of
two possible distinct states.
A bit is simply a 1 or a 0.
A true or a false.
It is the most basic unit of data in a computer. It is also called
machine language.
The two values can also be interpreted as logical values
(true/false, yes/no), algebraic signs (+/-), activation states
(on/off), or any other two valued attribute.
The length of a binary number may be referred to as its “bit-
length”.

2. BCD (Binary Coded Decimal)


BCD is a method of using binary digits to represent the decimal
digits 0 through 9.
A decimal digit is represented by four binary digits, as shown
below:

Decimal BCD NUMBER


Digits
0 0000
1 0001
2 0010
3 0011
4 0100
5 0101
6 0110
7 0111
8 1000
9 1001

BCD and binary are not the same. For example, 49 10 in binary is
1100012, but 4910 in BCD 01001001BCD. Each decimal digit is
converted to its binary equivalent.

BCD uses 4-bits to represent each decimal digit in a number, the


decimal range from 0 – 9.

BCD is a digital encoding method for numbers using decimal


notation, with each decimal digit represented by its own binary
sequence. In BCD, a numeral is usually represented by four-bits
which, in general, represent the decimal range 0 through 9 to
00000000 through 00001001.

For example,
DECIMAL NUMBER BCD NUMBER
5 0101
8 1000
105 0001 0000 0101
EXAMPLE: Determine the BCD value for the decimal number
5319.

SOLUTION: Since there are four digits in the decimal number,


there are four bytes in the BCD number. They are:
Thousands Hundreds Tens Units
5 3 1 9
00000101 00000011 00000001 00001001

3. ASCII
(America Standard Code for Information
Interchange)
The ASCII uses 7-bits for representation, that is, 2 7 = 128
characters. It is referred to as ASCII – 7. It is a 7-bit code with the
8th bit as the parity bit.
The ASCII character set, for example, uses the numbers 0 through
127 to represent all English characters as well as special control
characters.
The code was developed by the American National Standards
Institute (ANSI) to provide a standard code that could be used
by many different computer manufacturers in order to make them
compatible.
For example,
CHARACTERS ASCII – 7
5 011 0101
8 011 1000
9 011 1001
zone bits numeric bits
The American Standard Code for Information Interchange
(ASCII) is a character-encoding scheme based on the ordering of
the English alphabet. ASCII codes represent text in computers,
communications equipment, and other devices that use text. Most
modern character-encoding schemes are based on ASCII, though
they support many more characters than ASCII does.

4. EBCDIC
(Extended Binary Coded Decimal Interchange
Code)
It uses 8-bits coding system, that is, 28 = 256 character. It is
referred to as ASCII – 8, because it uses the full 8-bits.
For example,
CHARACTERS EBCDIC
5 1111 0101
8 1111 1000
9 1111 1001
Zone bits numeric bits

EBCDIC is an 8-bit character encoding used mainly on IBM


mainframe and IBM midrange computer operating systems.
EBCDIC was devised in 1963 and 1964 by IBM and was announced
with the release of the IBM System/360 line of mainframe
computers. It is an 8-bit character encoding, in contrast to, and
developed separately from, the 7-bit ASCII encoding scheme.
It was created to extend the existing Binary Coded Decimal (BCD)
interchange code, or BCDIC, which itself was devised as an
efficient means of encoding the two zone and number punches on
punched cards into 6-bits.

COMPUTER CHARACTER
SETS
Computer character sets is a defined list of characters
recognized by the computer hardware and software. It is the set of
characters represented in different data representation method.
Each character is represented by a number.

Character sets are basically the characters used by the


computers to represent certain data feed into it in a more
understandable manner to it.

A character denotes any letter, digit, punctuation symbols or any


other sign used in the representation of data or information in any
language.
The following character set:
1. Alphabetic character set, that is, letters (A-Z)
2. Numeric character set, that is, numbers or digits (0-9)
3. Alpha-Numeric character set, that is, (A –Z) and (0-9)
4. Boolean character set, that is, 1 or 0, true or false, male or
female etc.
5. Special symbols or characters e.g. /, +,), (, $, =, &, ? e.t.c.

The special characters that are regularly used in computing are as


follows:

+ Plus - Minus * Asterisk


/ Stroke / Slash = Equals $ Currency (dollar)
, Comma “ Quotation ( Open Bracket
) Close bracket > greater than < Less than
? Question mark & Ampersand ; Semicolon
‘ Apostrophe : Colon . Period
^ Caret % Per cent # Hash
@ At
THE STANDARD ASCII CHARACTER SET

Bytes are frequently used to hold individual characters in a text


document. In the ASCII character set, each binary value between 0
and 127 is given a specific character. Most computers extend the
ASCII character set to use the full range of 256 characters
available in a byte. The upper 128 characters handle special
things like accented characters from common foreign languages.

The ASCII character set, for example, uses the numbers 0


through 127 to represent all English characters as well as special
control characters.

Computers store text documents, both on disk and in memory,


using these codes.
Let us confirm this, open up a new file in Notepad and insert the
sentence, “ClickIT SS11 Computer” in it. Save the file to disk under
the name test.txt.
Then use the explorer and look at the size of the file. You will find
that the file has a size of 21bytes on disk: 1 byte for each
character.
If you add another word to the end of the sentence and re-save it,
the file size will jump to the appropriate number of bytes. Each
character consumes a byte.

TERMS USED IN DATA


REPRESENTATION

1. TYPECASTING:
This is the changing or converting one data type to another.

TYPES OF DATA TYPES OR FIELD TYPES:


i. Alphabetic data types or fields
ii. Numeric data types or fields
iii. Alpha-numeric data types or fields
iv. Date data types or fields
v. Boolean data types or fields

2. PARITY BIT
A parity bit is the addition of one bit to the end of the bit
representing a character. This is automatically added to the end of
the bits to make the total numbers of bits odd or even number
(Odd parity or Even parity)

3. PARITY CHECK OR CHECK DIGIT


Parity check or check digit is a means of ensuring that the

numbers of bits that makes up a character maintains its validity.

4. NIBBLE
A nibble or less commonly referred to as nybble is a computing

term for a four-bit aggregation, or half an octet (an octet being an

8-bit byte. As a nibble contains 4 bits, there are sixteen (2 4)

possible values, so a nibble corresponds to a single hexadecimal

digit. Thus, it is often referred to as a hex digit or hexit.

You might also like