Data Processing Work
Data Processing Work
PARTICIPANTS:
Mendoza Rosmer. IC 21.140.853
Introduction
Since ancient times, man has been processing data. It is very likely that
primitive man used his fingers to carry out very simple operations and store all the
possible information in his memory, so it was necessary to use all the means that
would allow him to solve slightly more complicated operations, such as the case of
the abacus which, in fact, was one of the first inventions. That is why data
processing is the Technique that consists of the collection of primary input data,
which are evaluated and organized, to obtain useful information, which will then be
analyzed by the end user, so that they can make decisions or carry out the actions
you deem appropriate.
The type of a data is the set of values that it can take during the program. If
you try to give it a value outside of the set, an error will occur.
Pascal is said to be a strongly typed language. This means that all data
must have an explicitly declared type, and also that there are certain restrictions on
expressions regarding the data types involved in them.
Data types, like almost all Pascal objects, can be declared. Type declaration
was already discussed in the topic corresponding to the structure of a program .
In Pascal there are a great variety and quantity of data types. But in this
tutorial we will only cover the basics so that you can build your first programs.
There are many classifications for data types, and depending on the source you
look at, they will show you one or the other. Below you have one of the possible
classifications.
Dynamic
Static
o The chain type
o Structured
o Simple
Ordinals
Non-ordinal
static type
Almost all data types are static, the exception is pointers and they will not be
discussed due to their complexity.
That a data type is static means that the size it occupies in memory cannot
vary during the execution of the program. That is, once Once a variable of a certain
type is declared , it is assigned a fixed chunk of memory, and this chunk cannot be
increased or decreased.
dynamic types
Only the pointer type falls into this category. This type allows you to have
greater control over memory management in your programs. With them you can
manage the size of your variables at run time, that is, when the program is running.
Simple types
As their name indicates, they are the basic types in Pascal. They are the
simplest and the easiest to learn. For all this, they will be the ones we focus on.
The most basic simple types are: integer , logical , character and real . And
most programming languages support them, unlike structured ones that can vary
from one language to another.
Structured types
Ordinal types
Within the simple types, the ordinal ones are the most abundant. A type is
said to be ordinal because the set of values it represents can be counted, that is,
we can establish a one-to-one relationship between its elements and the set of
natural numbers.
Non-ordinal types
Simplifying, we could reduce the simple non-ordinal types to the type real .
This type helps us declare variables that can take values within the set of real
numbers.
Meaning (semantics): What does it mean? From the meaning extracted from
information, each individual evaluates the possible consequences and adapts
their attitudes and actions in accordance with the foreseeable consequences
that are deduced from the meaning of the information. This refers to what rules
the individual or expert system must follow to modify their future expectations
about each possible alternative.
Importance (relative to the receiver): Does it address an important issue? The
importance of information for a recipient will refer to the degree to which it
changes the attitude or behavior of individuals. In modern societies, individuals
obtain a large amount of information from the mass media , a large part of it is
of little importance to them, because it alters the behavior of individuals in a
very insignificant way. This refers to to what quantitative degree future
expectations should be altered. Sometimes it is known that a fact makes some
things less probable and others more probable; the importance has to do with
how much less probable some alternatives will be compared to the others.
Validity (in the space-time dimension): Is it current or outdated? In practice, the
validity of information is difficult to evaluate, since in general accessing
information does not allow us to immediately know whether said information is
valid or not. This has to do with the synchronization in time of the signals that
allow expectations to be reassessed with the expectations at a given moment.
Validity (relative to the issuer): Is the issuer reliable or can it provide invalid
(false) information? This has to do with whether the cues should be considered
in the reevaluation of expectations or should be ignored because they are not
reliable cues.
Value (volatile intangible asset ): How useful is it to the recipient?
Data processing
Data processing is, generally, "the collection and manipulation of data elements to
produce meaningful information."
Data processing is a subset of information processing , "the changing (processing)
of information in any way detectable by an observer. Data processing is different
from text processing , as the latter manipulates texts instead of data.
Data Processing Functions
Data processing may involve various functions, including:
" Validation " - Ensure that the data provided is "clean, correct and useful."
Sorting - "Orders elements in a certain sequence and/or into different sets."
Recapitulation - reducing the details of the data to its main points.
Aggregation – combining multiple pieces of data.
" Analysis " - the "collection, organization, analysis, interpretation and
presentation of data."
Information - detailed list or summary of computerized information data.
This is the oldest process and involves the use of human resources, such as
performing mental calculations, recording data with pencil and paper, sorting
and classifying manually. This results in a slow and error-prone process
throughout all stages or activities of the processing cycle. Finally the results
are expressed in written form, creating large volumes of stored written
information. As an example, we can have the academic evaluation in a
school, where the teacher manually records, with a pen, the students'
evaluation notes in a table called registry, and then obtains the averages
through manual arithmetic calculations, leaving all this information stored in
your registry.
b. Mechanical Process
In this type of process, the link of information between the different elements
of information processing, storage and communication continues to be
carried out manually, but to carry out each of these tasks electromechanical
machines are used, with which we obtain greater efficiency. As an example,
we have the activities carried out with drilling machines, which, through the
use of electrical energy, will activate a motor, and the shaft of this motor with
its respective adapter will allow soils to be drilled at different depths. Other
examples: video camera, photographic camera, calculators, etc.
d. Electronic Processes
A brief example, suppose you sign up for some courses through the Internet,
where you must select the courses, the teachers and the schedules that interest
you, each time you enter a record into the system through the school's Web portal
[call it a course , subject or schedule] the database where that information resides
is modified in real time. Obviously you can enter, delete or edit records according
to your needs, this is done in real time and as long as you have the corresponding
permissions.
Input files: These files are made up of a series of data stored on an input device .
Output files: these files contain the information that is viewed from the computer.
Constant files: these files are made up of records whose fields are fixed or have a
low frequency of variation.
Historical files: are made up of data that vary over time and with information from
updated files.
Motion files: this type of files are used together with constants and have some
field in common.
Maneuver files: these are created when a program is executed and are deleted
once the execution ends, they are auxiliary.
Ascii: in this type of file the data is stored through simple text. This allows the data
they contain to be exchanged as well as to create files that the user himself can
modify.
Binary: this type of files, on the other hand, stores information in a language that
only the computer itself understands, for example colors , sounds, images or
commands. These files are smaller than the previous ones.
According to the mechanism used to access the files, the following are found:
Direct files: in these each record can be read and written directly. This is achieved
by expressing its address in the file.
Sequential files: in these files the records are organized sequentially in such a
way that to read one, the previous ones must have been read.
Index files: in these files the records can be accessed using the password they
have.
Records
In computing , or specifically in the context of a relational database , a
record (also called a row) represents a single object of data implicitly structured in
a table . In simple terms, a database table can be imagined as consisting of rows
and columns or fields . Each row in a table represents a set of related data, and all
rows in the same table have the same structure. A record is a set of fields that
contain data that belongs to the same entity repetition. A consecutive number
(record number) is automatically assigned, which is sometimes used as an index,
although it is normal and practical to assign each record a key field for searching.
Databases and types
A database or data bank is a set of data belonging to the same context and
systematically stored for later use. In this sense; A library can be considered a
database composed mostly of documents and texts printed on paper and indexed
for consultation. Currently, and due to the technological development of fields such
as computing and electronics , most databases are in digital format, this being an
electronic component, and therefore a wide range of solutions to the problem have
been developed and offered. of data storage.
There are programs called database management systems , DBMS for
short, that allow you to store and later access data in a quick and structured way.
The properties of these DBMSs, as well as their use and administration, are
studied within the field of computing.
The most common applications are for the management of companies and
public institutions; They are also widely used in scientific environments in order to
store experimental information.
Although databases can contain many types of data, some of them are
protected by the laws of various countries. For example in Spain, personal data is
protected by the Organic Law on the Protection of Personal Data (LOPD), in
Mexico by the Federal Law of Transparency and Access to Government Public
Information and in Argentina the Data Protection Law Personal . 1
Databases can be classified in several ways, according to the context being
managed, their usefulness or the needs they satisfy.
Depending on the variability of the database:
Static databases
They are read-only databases, used primarily to store historical data that can later
be used to study the behavior of a data set over time, make projections , make
decisions , and perform data analysis for business intelligence .
Dynamic databases
They are databases where the stored information is modified over time, allowing
operations such as updating, deleting and editing data, in addition to fundamental
query operations. An example may be the database used in a supermarket
information system.
According to the content:
Bibliographic databases
They only contain a surrogate (representative) of the primary source, which allows
it to be located. A typical record in a bibliographic database contains information
about the author, publication date, publisher, title, edition, of a given publication,
etc. It may contain a summary or extract of the original publication, but never the
full text, because otherwise we would be in the presence of a full-text database (or
primary sources - see below). As its name indicates, the content is figures or
numbers. For example, a collection of laboratory analysis results, among others.
Full text databases
They store primary sources, such as all the content of all editions of a collection of
scientific journals.
Directories
One example are the phones guides in electronic format.
Databases or "libraries" of chemical or biological information
They are databases that store different types of information from chemistry , life
sciences or medicine . They can be considered in several subtypes:
Those that store nucleotide or protein sequences.
The databases of metabolic pathways.
Structure databases, comprising records of experimental data on 3D structures
of biomolecules.
Clinical databases.
Bibliographic databases (biological, chemical, medical and other fields):
Puchen , Medline , EBSCOhost .
Database models
In addition to classification by database function, they can also be classified
according to their data management model.
A data model is basically a "description" of something known as a data
container (something where information is stored), as well as the methods for
storing and retrieving information from those containers. Data models are not
physical things: they are abstractions that allow the implementation of an efficient
database system; They generally refer to algorithms and mathematical concepts.
Some models frequently used in databases:
Hierarchical databases
Main article: Hierarchical database
In this model the data is organized in the form of an inverted tree (some say root),
where a parent information node can have several children. The node that has no
parents is called the root , and the nodes that have no children are known as
leaves.
Hierarchical databases are especially useful in the case of applications that
handle a large volume of information and highly shared data, allowing the creation
of stable and high-performance structures.
One of the main limitations of this model is its inability to efficiently represent data
redundancy.
Network database
Main article: Network database
This is a slightly different model from the hierarchical one; Its fundamental
difference is the modification of the node concept: the same node is allowed to
have several parents (a possibility not allowed in the hierarchical model).
It was a great improvement over the hierarchical model, as it offered an
efficient solution to the data redundancy problem; But, even so, the difficulty of
managing information in a network database has meant that it is a model used
mostly by programmers rather than end users.
Transactional databases
They are databases whose sole purpose is to send and receive data at high
speeds. These databases are very rare and are generally aimed at the quality
analysis environment, production and industrial data. It is important to understand
that their sole purpose is to collect and recover data at the highest speed possible,
therefore redundancy and duplication of information is not a problem as with other
databases, generally in order to make the most of them they allow some type of
connectivity to databases relational.
A common example of a transaction is the transfer of an amount of money
between bank accounts. It is normally carried out through two different operations,
one in which the balance of the source account is decreased and another in which
we increase the balance of the destination account. To guarantee the atomicity of
the system (that is, so that money does not appear or disappear), the two
operations must be atomic, that is, the system must guarantee that, under any
circumstances (even a system crash), the final result is that either the two
operations have been carried out, or neither has been carried out,
Relational databases
Main article: Relational model
This is the model currently used to represent real problems and manage
data dynamically. After its foundations were postulated in 1970 by Edgar Frank
Codd , from the IBM laboratories in San José (California) , it did not take long to
consolidate itself as a new paradigm in database models. Its fundamental idea is
the use of "relations". These relationships could logically be considered as sets of
data called " tuples ." Although this is the theory of relational databases created by
Codd, most of the time it is conceptualized in a way that is easier to imagine. This
is thinking of each relationship as if it were a table that is composed of records (the
rows of a table), which would represent the tuples, and fields (the columns of a
table).
In this model, where and how data is stored is irrelevant (unlike other
models such as hierarchical and network). This has the considerable advantage of
being easier for a casual user of the database to understand and use. Information
can be retrieved or stored using "queries" that offer extensive flexibility and power
to manage information.
The most common language to build queries to relational databases is
SQL , Structured Query Language, a standard implemented by the main engines
or relational database management systems.
During its design, a relational database goes through a process known as
database normalization .
During the 1980s, the appearance of dbase produced a revolution in
programming languages and data management systems. Although it should never
be forgotten that dbase did not use SQL as the base language for its management.
Multidimensional databases
Main article: Multidimensional database
They allow full-text indexing, and generally perform more powerful searches.
Tesauros is an index system optimized for this type of databases.
Deductive databases
A deductive database system is a database system but with the difference that it
allows deductions to be made through inferences. It is mainly based on rules and
facts that are stored in the database. Deductive databases are also called logical
databases, because they are based on mathematical logic. This type of database
arises due to the limitations of the Relational Database in responding to recursive
queries and deducing indirect relationships from the data stored in the database.
Language
It uses a subset of the Prolog language called Datalog which is declarative and
allows the computer to make inferences to answer queries based on stored facts
and rules.
Advantages
Ascending: where you start from the facts and obtain new ones by applying
inference rules.
Descending: where we start from the predicate (objective of the query carried
out) and try to find similarities between the variables that lead us to correct
facts stored in the database.
Distributed Database Management System (DBMS)
The database and DBMS software may be distributed across multiple sites
connected by a network. There are two types:
1. Homogeneous distributed: they use the same DBMS in multiple sites.
2. Heterogeneous distributed: Gives rise to federated DBMSs or multi-database
systems in which the participating DBMSs have a certain degree of local autonomy
and have access to several pre-existing autonomous databases stored in the
DBMSs, many of these employ a client-server.
These arise due to the physical existence of decentralized organizations. This
gives them the ability to join the databases of each location and thus access
different universities, store branches, among others.
Data processing center.
A data processing center is the set of physical, logical and human resources
necessary for the organization, execution and control of a company's computing
activities.
You can find various names referring to the data processing center, such as IT
department, data processing department, IT center, computer room or calculation
center.
Functions of a CPD
Within a company, it fulfills various functions that justify the established jobs that
exist in it, these functions are the following:
• Development of computer systems
Location
Among the most important factors that motivate the creation of a CPD, we can
highlight guaranteeing the continuity of service to clients, employees, citizens,
suppliers and collaborating companies, since in these areas the physical protection
of computer or communications equipment is very important. involved, as well as
database servers that may contain critical information.
Design
The design of a data processing center begins with the choice of its geographical
location , and requires a balance between various factors:
In a data processing center there are various jobs for the different areas and
functions carried out there. The classification presented below cannot be rigorous
due to the many nuances that occur in IT personnel, depending on the type of
company, size of the data processing center and the type and volume of
applications carried out.
Project Manager Reporting directly to the head of the development area, his
mission consists of directing a computer project from the specifications and needs
of users to its exploitation.
Analysts Based on user requirements and under the orders of the project
manager, they must carry out the analysis of the applications and help the
programmers in their development.
Programmers They receive the analysis of the applications from the analysts and,
based on it, they design the diagram or pseudocode, encoding it in the chosen
language, and are also responsible for creating the documentation aimed at the
user.
Operators They are responsible for the direct operation and operation of the
system, execution of processes, preparation of supports, peripherals and
necessary materials.
Conclusion