0% found this document useful (0 votes)
89 views

Assignment Data

The document discusses data editing, processing, classification, and presentation. It defines data editing as reviewing collected survey data to detect and correct errors. Data processing involves collecting, preparing, inputting, processing, outputting, and storing data. Classification involves arranging data into groups based on attributes or numerical characteristics. Common ways to present data include pie charts, line charts, histograms, and two-dimensional diagrams.

Uploaded by

shyam tomer
Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
89 views

Assignment Data

The document discusses data editing, processing, classification, and presentation. It defines data editing as reviewing collected survey data to detect and correct errors. Data processing involves collecting, preparing, inputting, processing, outputting, and storing data. Classification involves arranging data into groups based on attributes or numerical characteristics. Common ways to present data include pie charts, line charts, histograms, and two-dimensional diagrams.

Uploaded by

shyam tomer
Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 7

Gautam Buddha University

Department of social work

Bachelors of Social Work

Data Editing, Processing and presentation

Submitted To:- Submitted By:-

Dr. Ronak Ahmed Samiksha Patel

(18/UBSW/010)
Contents

1. Data Editing
2. Classification of Data
3. Data Processing
4. Data Presentation
Data editing

Data editing is defined as the process involving the review and adjustment of collected
survey data. The purpose is to control the quality of the collected data. Data editing can
be performed manually, with the assistance of a computer or a combination of both.

It is a process of examining the raw data to detect errors and omissions and to correct
them, if possible, so as to ensure legibility, completeness, consistency and accuracy.
The recorded data must be legible so that it could he coded later. The inaccuracy of the
survey data may be due to interviewer bias or cheating. One way of spotting is to look
for a common pattern of responses in the instrument of a particular interviewer.

The editing can be done at two stages:

1. Field Editing

2. Central Editing.

Field Editing: The field editing consists of review of the reporting forms by the
investigator for completing or translating what the latter has written in abbreviated form
at the time of interviewing the respondent. This form of editing is necessary in view of
the writing of individuals, which vary from individual to individual and sometimes difficult
for the tabulator to understand. This sort of editing should be done as soon as possible
after the interview, as it may be necessary sometimes to recall the memory. While doing
so, care should be taken so that the investigator does not correct the errors of omission
by simply guessing what the respondent would have answered if the question was put
to him.

Central Editing: Central editing should be carried out when all the forms of schedules
have been completed and returned to the headquarters. This type of editing requires
that all the forms are thoroughly edited by a single person (editor) in a small field study
or a small group of persons in case of a large field study, The editor may correct the
obvious errors, such as an entry in a wrong place, entry recorded in daily terms
whereas it should have been recorded in weeks/months, etc. Sometimes, inappropriate
or missing replies can also be recorded by the editor by reviewing the other information
recorded in the schedule. If necessary, the respondent may be contacted for
clarification. All the incorrect replies, which are quite obvious, must be deleted from the
schedules.

CLASSIFICATION OF DATA

It is the process of arranging data in groups or classes on the basis of some


characteristics. Classification condenses the data, facilitates comparisons, helps to
study the relationships and facilitates in statistical treatment of data. The classification
should be unambiguous and mutually exclusive and collectively exhaustive. It should
not only be flexible but also suitable for the purpose for which it is sought.' Classification
can either be according to attributes or numerical characteristics.

1. Classification According to Attributes: To classify the data according to attributes


we use descriptive characteristics like sex, caste, education, user of a product etc. The
descriptive characters are the one which cannot be measured quantitatively. One can
only talk in terms of its presence or absence. The classification according to attributes
may be of two types.

i) Simple Classification: In the case of simple classification each class is divided into
two sub classes and only one attribute is studied viz, user of a product or non-user of a
product, married or unmarried, employed or unemployed, Brahmin or non-Brahmin etc.

ii) Manifold Classification: In the case of manifold classification more than one
attributes are considered. For example, the respondents in a survey may be classified
as user of a particular brand of a product and non-user of particular brand of product.
Both user and non-user can be further classified into male and female.

2. Classification According to Numerical Characteristic: When the observations


possesses numerical characteristics such as sales, profits, height, weight, income,
marks, they are classified according to class intervals.
Data processing

Data processing occurs when data is collected and translated into usable information.
Usually performed by a data scientist or team of data scientists, it is important for data
processing to be done correctly as not to negatively affect the end product, or data
output.

Data processing starts with data in its raw form and converts it into a more readable
format (graphs, documents, etc.), giving it the form and context necessary to be
interpreted by computers and utilized by employees throughout an organization.

Six stages of data processing

1. Data collection

Collecting data is the first step in data processing. Data is pulled from available sources,
including data lakes and data warehouses. It is important that the data sources
available are trustworthy and well-built so the data collected is of the highest possible
quality.

2. Data preparation

Once the data is collected, it then enters the data preparation stage. Data preparation,
often referred to as “pre-processing” is the stage at which raw data is cleaned up and
organized for the following stage of data processing. During preparation, raw data is
diligently checked for any errors. The purpose of this step is to eliminate bad data and
begin to create high-quality data for the best business intelligence.

3. Data input

The clean data is then entered into its destination and translated into a language that it
can understand. Data input is the first stage in which raw data begins to take the form of
usable information.
4. Processing

During this stage, the data inputted to the computer in the previous stage is actually
processed for interpretation. Processing is done using machine learning algorithms,
though the process itself may vary slightly depending on the source of data being
processed (data lakes, social networks, connected devices etc.) and its intended use
(examining advertising patterns, medical diagnosis from connected devices,
determining customer needs, etc.).

5. Data output/interpretation

The output/interpretation stage is the stage at which data is finally usable to non-data
scientists. It is translated, readable, and often in the form of graphs, videos, images,
plain text, etc.). Members of the company or institution can now begin to self-serve the
data for their own data analytics projects.

6. Data storage

The final stage of data processing is storage. After all of the data is processed, it is then
stored for future use. While some information may be put to use immediately, much of it
will serve a purpose later on. Plus, properly stored data is a necessity for compliance
with data protection legislation like GDPR. When data is properly stored, it can be
quickly and easily accessed by members of the organization when needed.

PRESENTATION OF DATA

The presentation of data by different ways are:

1. Pie Chart
2. Line chart
3. Histogram
4. Two dimensional diagram-s

Pie Chart

In the pie chart there are different segments of a circle represent percentage
contribution of various components to the total. It brings out the relative importance of
various components of data. For drawing a pie chart, we construct a circle of any
diameter and then the circle is broken into a desired number of segments, angle 360
degree representing 100 percent of data through which the total data will be
represented.

Line chart

The line chart helps to identify the trends of data whether it is going to the upward trend
or downward trend. The best example of line chart is the price of a any stock in the
market.For example, price of any product, sales, import and export data.

Histogram

A histogram is a plot that lets you discover, and show, the underlying frequency
distribution (shape) of a set of continues data. This allows the inspection of the data for
its underlying distribution outliers, skewness, etc

Two dimensional diagrams

The commonly used two dimensional diagrams are rectangular diagrams and squares.
In rectangular diagrams the rectangles are used to present the data in the graphic form.
These diagrams are used for comparing two sets of data. The height of the rectangle is
proportional to the ratio of the data which bear to each other in a given series and the
width of the rectangle varies in proportion to the aggregate. Before constructing the
rectangular diagram, the data is converted into percentages.

You might also like