Project Title: Hospital Management System Description: Hospital Management System Is A Web Based Project. It Consists of Client Registration
Project Title: Hospital Management System Description: Hospital Management System Is A Web Based Project. It Consists of Client Registration
DESCRIPTION:
Client registration module consists of doctors and hospital information were doctor maintains
web sites to help doctors practices run more efficiently. These sites are easing the staff workload,
decreasing phone traffic, and helping to improve patient care and hospital maintains telephone
traffic in offices increases as patients call to find out if flu shots are available, and when the
hospitals will be scheduled. This is difficult for staff, who often are tied up responding to these
telephone calls rather than doing their regular work.
Patient registration module consists of fixing an appointment for doctor maintains once patients
go to the Internet for information about your practice, you can use your Web site as a tool to
guide them to important information. When I find a new link or post information that patients
may find useful, I indicate the new material on my home page... and fixing an appointment on
hospital maintains You can post the relevant information, along with the times and dates of flu
shot hospitals, on the Web site as soon as the flu vaccines arrive. The best time to remind
patients about the site is during a clinical visit or when they are checking in or out.
1. INTRODUCTION
2. SYSTEM ANALYSIS
2.1. INTRODUCTION
2.2. SYSTEM WORKFLOW
2.3. STUDY OF THE SYSTEM
2.4. HARDWARE & SOFTWARE REQUIRMENT
2.5. PROPOSED SYSTEM
2.6. INPUT & OUTPUT
2.7. PROCESS MODELS USED WITH JUSTIFICATION
3. FEASIBILITY REPORT
5. SELECTED SOFTWARE
6. SYSTEM DESIGN
6.1. INTRODUCTION
6.2. SYSTEM WORKFLOW
6.3. NORMALIZATION
6.4. E-R DIAGRAM
6.5. DATA FLOW DIAGRAMS
6.6. DATA DICTIONARY
6.7. UML DIAGRAMS
7. OUTPUT SCREENS
8.1. INTRODUCTION
8.2. STRATEGIC APPROACH OF SOFTWARE TESTING
8.3. UNIT TESTING
9. SYSTEM SECURITY
9.1. INTRODUCTION
10. CONCLUSION
12. BIBLIOGRAPHY
INTRODUCTION
INTRODUCTION TO PROJECT
The best-known type of hospital is the general hospital, which is set up to deal with many kinds
of disease and injury, and typically has anemergency department to deal with immediate and
urgent threats to health. A general hospital typically is the major health care facility in its region,
with large numbers of beds for intensive care and long-term care; and specialized facilities
for surgery, plastic surgery, childbirth, bioassay laboratories, and so forth. Larger cities may have
several hospitals of varying sizes and facilities. Some hospitals, especially in the United States,
have their own ambulance service.
Existing system:
In the present existing system the drawback is that the patient need to take the appointment
manually from the doctor, while he is going for the medical checkup this process is time taking
for the common person in this busy generation to overcome this we are going have a solution
through online procedure.
SOFTWARE SOLUTIONS
SPEED:
EXPERTISE:
Our teams combine cutting edge technology skills with rich domain expertise.
What’s equally important - they share a strong customer orientation that means
they actually start by listening to the customer. They’re focused on coming up with
solutions that serve customer requirements today and anticipate future needs.
They offer customers the advantage of being able to Architect, integrate and
manage technology services. This means that they can rely on one, fully
accountable source instead of trying to integrate disparate multi vendor solutions.
SERVICES:
A hospital information system (HIS), variously also called clinical information system (CIS) is a
comprehensive, integrated information system designed to manage the administrative, financial and
clinical aspects of a hospital. This encompasses paper-based information processing as well as data
processing machines.
It can be composed of one or a few software components with specialty-specific extensions as well as of
a large variety of sub-systems in medical specialties (e.g. Laboratory Information System, Radiology
Information System).
CISs are sometimes separated from HISs in that the former concentrate on patient-related and clinical-
state-related data (electronic patient record) whereas the latter keeps track of administrative issues. The
distinction is not always clear and there is contradictory evidence against a consistent use of both terms.
SYSTEM ANALYSIS
2.1. INTRODUCTION
After analyzing the requirements of the task to be performed, the next step is to
analyze the problem and understand its context. The first activity in the phase is
studying the existing system and other is to understand the requirements and
domain of the new system. Both the activities are equally important, but the first
activity serves as a basis of giving the functional specifications and then successful
design of the proposed system. Understanding the properties and requirements of a
new system is more difficult and requires creative thinking and understanding of
existing running system is also difficult, improper understanding of present system
can lead diversion from solution.
This document play a vital role in the development of life cycle (SDLC) as it
describes the complete requirement of the system. It means for use by
developers and will be the basic during testing phase. Any changes made to the
requirements in the future will have to go through formal change approval
process.
SPIRAL MODEL was defined by Barry Boehm in his 1988 article, “A spiral
Model of Software Development and Enhancement. This model was not the first
model to discuss iterative development, but it was the first model to explain why
the iteration models.
GUI’S
In the flexibility of the uses the interface has been developed a graphics concept in
mind, associated through a browses interface. The GUI’S at the top level have been
categorized as
The operational or generic user interface helps the users upon the system in
transactions through the existing data and required services. The operational user
interface also helps the ordinary users in managing their own information helps the
ordinary users in managing their own information in a customized manner as per
the assisted flexibilities.
NUMBER OF MODULES
The system after careful analysis has been identified to be presented with the
following modules:
The major module in the project is admin total things can managed by admin only.
1.Doctors
2.Hospital
3.Patient
Doctors:
Doctors can give appointment to Hospital, i.e available time in the hospital, doctor
itself register in hospital for authentication and get login details from the admin in
the hospital and check for appointments for the patients
Hospital:
Itself register, and provide facilities to the doctors and patients maintain the doctors
and patient list. If necessary manage and create list for doctors and patients.
Patient;
Can wait for doctor’s appointment and choose hospital for specialist doctors.
Patient will donate blood or receive from other people.
Authentication:-
This module contains all the information about the authenticated user. User without
his username and password can’t enter into the login if he is only the authenticated
user then he can enter to his login and he can see the quotation and give the
quotation for the particular products
P 4 ( 733 MHZ )
512 RAM
16 Bit High Colors
Software Requirements:
The entire scope has been classified into five streams known as: -
Coordinator Level
Management Level
Auditor Level
User Level
INPUTS:
OUTPUTS:
The following commands specify access control identifiers and they are typically
used to authorize and authenticate the user (command codes are shown in
parentheses)
PASSWORD (PASS)
This command must be immediately preceded by the user name command, and, for
some sites, completes the user's identification for access control. Since password
information is quite sensitive, it is desirable in general to "mask" it or suppress type
out.
Chapter 3
Feasibility Report
Preliminary investigation examine project feasibility, the likelihood the system will be
useful to the organization. The main objective of the feasibility study is to test the
Technical, Operational and Economical feasibility for adding new modules and debugging old
running system. All system is feasible if they are unlimited resources and infinite time.
There are aspects in the feasibility study portion of the preliminary investigation:
• Technical Feasibility
• Operation Feasibility
• Economical Feasibility
The technical issue usually raised during the feasibility stage of the investigation
includes the following:
Proposed projects are beneficial only if they can be turned out into information
system. That will meet the organization’s operating requirements. Operational feasibility
aspects of the project are to be taken as an important part of the project implementation.
Some of the important issues raised are to test the operational feasibility of a project
includes the following: -
The well-planned design would ensure the optimal utilization of the computer resources and
would help in the improvement of performance status.
A system can be developed technically and that will be used if installed must still be a good
investment for the organization. In the economical feasibility, the development cost in
creating the system is evaluated against the ultimate benefit derived from the new systems.
Financial benefits must equal or exceed the costs.
The system is economically feasible. It does not require any addition hardware or
software. Since the interface for this system is developed using the existing resources and
technologies available at NIC, There is nominal expenditure and economical feasibility for
certain.
Chapter 4
SOFTWARE REQUIREMENT
SPECIFICATION
The software, Site Explorer is designed for management of web sites from a
remote location.
INTRODUCTION
Purpose: The main purpose for preparing this document is to give a general insight
into the analysis and requirements of the existing system or situation and for
determining the operating characteristics of the system.
Scope: This Document plays a vital role in the development life cycle (SDLC) and it
describes the complete requirement of the system. It is meant for use by the
developers and will be the basic during testing phase. Any changes made to the
requirements in the future will have to go through formal change approval process.
• Developing the system, which meets the SRS and solving all the
requirements of the system?
• Demonstrating the system and installing the system at client's location after
the acceptance testing is successful.
• Submitting the required user manual describing the system interfaces to
work on it and also the documents of the system.
• Conducting any user training that might be needed for using the system.
• Maintaining the system for a period of one year after installation.
OUTPUT DESIGN
OUTPUT DEFINITION
For Example
Output Media:
The outputs were needed to be generated as a hot copy and as well as queries to
be viewed on the screen. Keeping in view these outputs, the format for the output
is taken from the outputs, which are currently being obtained after manual
processing. The standard printer is to be used as output media for hard copies.
INPUT DESIGN
Input design is a part of overall system design. The main objective during the input
design is as given below:
INPUT STAGES:
• Data recording
• Data transcription
• Data conversion
• Data verification
• Data control
• Data transmission
• Data validation
• Data correction
INPUT TYPES:
INPUT MEDIA:
At this stage choice has to be made about the input media. To conclude
about the input media consideration has to be given to;
• Type of input
• Flexibility of format
• Speed
• Accuracy
• Verification methods
• Rejection rates
• Ease of correction
• Storage and handling requirements
• Security
• Easy to use
• Portability
Keeping in view the above description of the input types and input media, it
can be said that most of the inputs are of the form of internal and interactive. As
Input data is to be the directly keyed in by the user, the keyboard can be
considered to be the most suitable input device.
ERROR AVOIDANCE
At this stage care is to be taken to ensure that input data remains accurate
form the stage at which it is recorded upto the stage in which the data is accepted
by the system. This can be achieved only by means of careful control each time the
data is handled.
ERROR DETECTION
Even though every effort is make to avoid the occurrence of errors, still a
small proportion of errors is always likely to occur, these types of errors can be
discovered by using validations to check the input data.
DATA VALIDATION
The system is designed to be a user friendly one. In other words the system
has been designed to communicate effectively with the user. The system has been
designed with pop up menus.
It is essential to consult the system users and discuss their needs while
designing the user interface:
In the computer initiated interfaces the computer guides the progress of the
user/computer dialogue. Information is displayed and the user response of the
computer takes action or displays further information.
USER_INITIATED INTERGFACES
COMPUTER-INITIATED INTERFACES
Right from the start the system is going to be menu driven, the opening
menu displays the available options. Choosing one option gives another popup
menu with more options. In this way every option leads the users to data entry
form where the user can key in the data.
The requirement specification for any system can be broadly stated as given below:
The existing system is completely dependent on the user to perform all the duties.
Chapter 5
SELECTED SOFTWARE
5.1. INTRODUCTION TO .NET Framework
Generate managed code that conforms to the CTS. This means that managed
code can consume other managed types and instances, while strictly enforcing type
fidelity and type safety.
In addition, the managed environment of the runtime eliminates many
common software issues. For example, the runtime automatically handles object
layout and manages references to objects, releasing them when they are no longer
being used. This automatic memory management resolves the two most common
application errors, memory leaks and invalid memory references.
The runtime also accelerates developer productivity. For example,
programmers can write applications in their development language of choice, yet
take full advantage of the runtime, the class library, and components written in
other languages by other developers. Any compiler vendor who chooses to target
the runtime can do so. Language compilers that target the .NET Framework make
the features of the .NET Framework available to existing code written in that
language, greatly easing the migration process for existing applications.
While the runtime is designed for the software of the future, it also supports
software of today and yesterday. Interoperability between managed and
unmanaged code enables developers to continue to use necessary COM
components and DLLs.
The runtime is designed to enhance performance. Although the common
language runtime provides many standard runtime services, managed code is never
interpreted. A feature called just-in-time (JIT) compiling enables all managed code
to run in the native machine language of the system on which it is executing.
Meanwhile, the memory manager removes the possibilities of fragmented memory
and increases memory locality-of-reference to further increase performance.
Finally, the runtime can be hosted by high-performance, server-side
applications, such as Microsoft® SQL Server™ and Internet Information Services
(IIS). This infrastructure enables you to use managed code to write your business
logic, while still enjoying the superior performance of the industry's best enterprise
servers that support runtime hosting.
The .NET Framework class library is a collection of reusable types that tightly
integrate with the common language runtime. The class library is object oriented,
providing types from which your own managed code can derive functionality. This
not only makes the .NET Framework types easy to use, but also reduces the time
associated with learning new features of the .NET Framework. In addition, third-
party components can integrate seamlessly with classes in the .NET Framework.
For example, the .NET Framework collection classes implement a set of
interfaces that you can use to develop your own collection classes. Your collection
classes will blend seamlessly with the classes in the .NET Framework.
As you would expect from an object-oriented class library, the .NET
Framework types enable you to accomplish a range of common programming tasks,
including tasks such as string management, data collection, database connectivity,
and file access. In addition to these common tasks, the class library includes types
that support a variety of specialized development scenarios. For example, you can
use the .NET Framework to develop the following types of applications and services:
• Console applications.
• Scripted or hosted applications.
• Windows GUI applications (Windows Forms).
• ASP.NET applications.
• XML Web services.
• Windows services.
For example, the Windows Forms classes are a comprehensive set of
reusable types that vastly simplify Windows GUI development. If you write an
ASP.NET Web Form application, you can use the Web Forms classes.
ASP.NET
Server Application Development
Server-side applications in the managed world are implemented through
runtime hosts. Unmanaged applications host the common language runtime, which
allows your custom managed code to control the behavior of the server. This model
provides you with all the features of the common language runtime and class
library while gaining the performance and scalability of the host server.
The following illustration shows a basic network schema with managed code
running in different server environments. Servers such as IIS and SQL Server can
perform standard operations while your application logic executes through the
managed code.
ASP.NET is the hosting environment that enables developers to use the .NET
Framework to target Web-based applications. However, ASP.NET is more than just
a runtime host; it is a complete architecture for developing Web sites and Internet-
distributed objects using managed code. Both Web Forms and XML Web services
use IIS and ASP.NET as the publishing mechanism for applications, and both have a
collection of supporting classes in the .NET Framework.
XML Web services, an important evolution in Web-based technology, are
distributed, server-side application components similar to common Web sites.
However, unlike Web-based applications, XML Web services components have no UI
and are not targeted for browsers such as Internet Explorer and Netscape
Navigator. Instead, XML Web services consist of reusable software components
designed to be consumed by other applications, such as traditional client
applications, Web-based applications, or even other XML Web services. As a result,
XML Web services technology is rapidly moving application development and
deployment into the highly distributed environment of the Internet.
If you have used earlier versions of ASP technology, you will immediately
notice the improvements that ASP.NET and Web Forms offers. For example, you
can develop Web Forms pages in any language that supports the .NET Framework.
In addition, your code no longer needs to share the same file with your HTTP text
(although it can continue to do so if you prefer). Web Forms pages execute in
native machine language because, like any other managed application, they take
full advantage of the runtime. In contrast, unmanaged ASP pages are always
scripted and interpreted. ASP.NET pages are faster, more functional, and easier to
develop than unmanaged ASP pages because they interact with the runtime like
any managed application.
The .NET Framework also provides a collection of classes and tools to aid in
development and consumption of XML Web services applications. XML Web services
are built on standards such as SOAP (a remote procedure-call protocol), XML (an
extensible data format), and WSDL ( the Web Services Description Language).
The .NET Framework is built on these standards to promote interoperability with
non-Microsoft solutions.
For example, the Web Services Description Language tool included with
the .NET Framework SDK can query an XML Web service published on the Web,
parse its WSDL description, and produce C# or Visual Basic source code that your
application can use to become a client of the XML Web service. The source code can
create classes derived from classes in the class library that handle all the underlying
communication using SOAP and XML parsing. Although you can use the class library
to consume XML Web services directly, the Web Services Description Language tool
and the other tools contained in the SDK facilitate your development efforts with
the .NET Framework.
If you develop and publish your own XML Web service, the .NET Framework
provides a set of classes that conform to all the underlying communication
standards, such as SOAP, WSDL, and XML. Using those classes enables you to focus
on the logic of your service, without concerning yourself with the communications
infrastructure required by distributed software development.
Finally, like Web Forms pages in the managed environment, your XML Web service
will run with the speed of native machine language using the scalable
communication of IIS.
LANGUAGE SUPPORT
The Microsoft .NET Platform currently offers built-in support for three
languages: C#, Visual Basic, and JScript.
ASP.NET Web Forms pages are text files with an .aspx file name extension.
They can be deployed throughout an IIS virtual root directory tree. When a browser
client requests .aspx resources, the ASP.NET runtime parses and compiles the
target file into a .NET Framework class. This class can then be used to dynamically
process incoming requests. (Note that the .aspx file is compiled only the first time it
is accessed; the compiled type instance is then reused across multiple requests).
An ASP.NET page can be created simply by taking an existing HTML file and
changing its file name extension to .aspx (no modification of code is required). For
example, the following sample demonstrates a simple HTML page that collects a
user's name and category preference and then performs a form postback to the
originating page when a button is clicked:
ASP.NET provides syntax compatibility with existing ASP pages. This includes
support for <% %> code render blocks that can be intermixed with HTML content
within an .aspx file. These code blocks execute in a top-down manner at page
render time.
ADO.NET OVERVIEW
Connections:
Connections are used to 'talk to' databases, and are represented by provider-
specific classes such as SqlConnection. Commands travel over connections and
resultsets are returned in the form of streams which can be read by a DataReader
object, or pushed into a DataSet object.
Commands:
Commands contain the information that is submitted to a database, and are
represented by provider-specific classes such as SqlCommand. A command can be
a stored procedure call, an UPDATE statement, or a statement that returns results.
You can also use input and output parameters, and return values as part of your
command syntax. The example below shows how to issue an INSERT statement
against the Northwind database.
DataReaders:
The DataReader object is somewhat synonymous with a read-only/forward-only
cursor over data. The DataReader API supports flat as well as hierarchical data. A
DataReader object is returned after executing a command against a database. The
format of the returned DataReader object is different from a recordset. For
example, you might use the DataReader to show the results of a search list in a
web page.
DATAADAPTERS (OLEDB/SQL)
The DataAdapter object works as a bridge between the DataSet and the
source data. Using the provider-specific SqlDataAdapter (along with its associated
SqlCommand and SqlConnection) can increase overall performance when
working with a Microsoft SQL Server databases. For other OLE DB-supported
databases, you would use the OleDbDataAdapter object and its associated
OleDbCommand and OleDbConnection objects.
The DataAdapter object uses commands to update the data source after
changes have been made to the DataSet. Using the Fill method of the
DataAdapter calls the SELECT command; using the Update method calls the
INSERT, UPDATE or DELETE command for each changed row. You can explicitly set
these commands in order to control the statements used at runtime to resolve
changes, including the use of stored procedures. For ad-hoc scenarios, a
CommandBuilder object can generate these at run-time based upon a select
statement. However, this run-time generation requires an extra round-trip to the
server in order to gather required metadata, so explicitly providing the INSERT,
UPDATE, and DELETE commands at design time will result in better run-time
performance.
1. ADO.NET is the next evolution of ADO for the .Net Framework.
2. ADO.NET was created with n-Tier, statelessness and XML in the forefront.
Two new objects, the DataSet and DataAdapter, are provided for these
scenarios.
3. ADO.NET can be used to get data from a stream, or to store data in a cache
for updates.
4. There is a lot more information about ADO.NET in the documentation.
5. Remember, you can execute a command directly against the database in
order to do inserts, updates, and deletes. You don't need to first put data into a
DataSet in order to insert, update, or delete it.
6. Also, you can use a DataSet to bind to the data, move through the data, and
navigate data relationships
SQL SERVER
A database management, or DBMS, gives the user access to their data and
helps them transform the data into information. Such database management
systems include dBase, paradox, IMS, SQL Server and SQL Server. These systems
allow users to create, update and extract information from their database.
During an SQL Server Database design project, the analysis of your business
needs identifies all the fields or attributes of interest. If your business needs
change over time, you define any additional fields or change the definition of
existing fields.
SQL Server stores records relating to each other in a table. Different tables
are created for the various groups of information. Related tables are grouped
together to form a database.
PRIMARY KEY
Every table in SQL Server has a field or a combination of fields that uniquely
identifies each record in the table. The Unique identifier is called the Primary Key,
or simply the Key. The primary key provides the means to distinguish one record
from all other in a table. It allows the user and the database system to identify,
locate and refer to one particular record in the database.
RELATIONAL DATABASE
FOREIGN KEY
When a field is one table matches the primary key of another field is referred
to as a foreign key. A foreign key is a field or a group of fields in one table whose
values match those of the primary key of another table.
REFERENTIAL INTEGRITY
Not only does SQL Server allow you to link multiple tables, it also maintains
consistency between them. Ensuring that the data among related tables is
correctly matched is referred to as maintaining referential integrity.
DATA ABSTRACTION
A major purpose of a database system is to provide users with an abstract
view of the data. This system hides certain details of how the data is stored and
maintained. Data abstraction is divided into three levels.
Physical level: This is the lowest level of abstraction at which one describes how
the data are actually stored.
Conceptual Level: At this level of database abstraction all the attributed and
what data are actually stored is described and entries and relationship among
them.
View level: This is the highest level of abstraction at which one describes only
part of the database.
ADVANTAGES OF RDBMS
DISADVANTAGES OF DBMS
SQL SERVER is a truly portable, distributed, and open DBMS that delivers
unmatched performance, continuous operation and support for every database.
SQL SERVER RDBMS is high performance fault tolerant DBMS which is specially
designed for online transactions processing and for handling large database
application.
SQL SERVER with transactions processing option offers two features which
contribute to very high level of transaction processing throughput, which are
The unrivaled portability and connectivity of the SQL SERVER DBMS enables
all the systems in the organization to be linked into a singular, integrated
computing resource.
PORTABILITY
OPEN SYSTEMS
UNMATCHED PERFORMANCE
The most advanced architecture in the industry allows the SQL SERVER
DBMS to deliver unmatched performance.
NO I/O BOTTLENECKS
SQL Server’s fast commit groups commit and deferred write technologies
dramatically reduce disk I/O bottlenecks. While some database write whole data
block to disk at commit time, SQL Server commits transactions with at most
sequential log file on disk at commit time, On high throughput systems, one
sequential writes typically group commit multiple transactions. Data read by the
transaction remains as shared memory so that other transactions may access that
data without reading it again from disk. Since fast commits write all data necessary
to the recovery to the log file, modified blocks are written back to the database
independently of the transaction commit, when written from memory to disk.
Chapter 6
SYSTEM DESIGN
6.1. INTRODUCTION
Software design sits at the technical kernel of the software engineering
process and is applied regardless of the development paradigm and area of
application. Design is the first step in the development phase for any engineered
product or system. The designer’s goal is to produce a model or representation of
an entity that will later be built. Beginning, once system requirement have been
specified and analyzed, system design is the first of the three technical activities
-design, code and test that is required to build and verify software.
The importance can be stated with a single word “Quality”. Design is the
place where quality is fostered in software development. Design provides us with
representations of software that can assess for quality. Design is the only way that
we can accurately translate a customer’s view into a finished software product or
system. Software design serves as a foundation for all the software engineering
steps that follow. Without a strong design we risk building an unstable system –
one that will be difficult to test, one whose quality cannot be assessed until the last
stage.
During design, progressive refinement of data structure, program structure,
and procedural details are developed reviewed and documented. System design can
be viewed from either technical or project management perspective. From the
technical point of view, design is comprised of four activities – architectural design,
data structure design, interface design and procedural design.
Insertion anomaly: Inability to add data to the database due to absence of other
data.
Normal Forms: These are the rules for structuring relations that eliminate
anomalies.
A relation is said to be in first normal form if the values in the relation are
atomic for every attribute in the relation. By this we mean simply that no attribute
value can be a set of values or, as it is sometimes expressed, a repeating group.
Transitive Dependency: If two non key attributes depend on each other as well
as on the primary key then they are said to be transitively dependent.
The above normalization principles were applied to decompose the data in
multiple tables thereby making the data to be maintained in a consistent state.
6.4. E – R DIAGRAMS
• The set of primary components that are identified by the ERD are
The primary purpose of the ERD is to represent data objects and their
relationships.
A data flow diagram is graphical tool used to describe and analyze movement
of data through a system. These are the central tool and the basis from which the
other components are developed. The transformation of data from input to output,
through processed, may be described logically and independently of physical
components associated with the system. These are known as the logical data flow
diagrams. The physical data flow diagrams show the actual implements and
movement of data between people, departments and workstations. A full
description of a system actually consists of a set of data flow diagrams. Using two
familiar notations Yourdon, Gane and Sarson notation develops the data flow
diagrams. Each component in a DFD is labeled with a descriptive name. Process is
further identified with a number that will be used for identification purpose. The
development of DFD’S is done in several levels. Each process in lower level
diagrams can be broken down into a more detailed DFD in the next level. The lop-
level diagram is often called context diagram. It consists a single process bit, which
plays vital role in studying the current system. The process in the context level
diagram is exploded into other process at the first level DFD.
The idea behind the explosion of a process into more process is that
understanding at one level of detail is exploded into greater detail at the next level.
This is done until further explosion is necessary and an adequate amount of detail is
described for analyst to understand the process.
A DFD is also known as a “bubble Chart” has the purpose of clarifying system
requirements and identifying major transformations that will become programs in
system design. So it is the starting point of the design to the lowest level of detail.
A DFD consists of a series of bubbles joined by data flows in the system.
DFD SYMBOLS:
Data flow
Data Store
CONSTRUCTING A DFD:
1. Process should be named and numbered for an easy reference. Each name
should be representative of the process.
2. The direction of flow is from top to bottom and from left to right. Data
traditionally flow from source to the destination although they may flow back to
the source. One way to indicate this is to draw long flow line back to a source.
An alternative way is to repeat the source symbol as a destination. Since it is
used more than once in the DFD it is marked with a short diagonal.
3. When a process is exploded into lower level details, they are numbered.
4. The names of data stores and destinations are written in capital letters. Process
and dataflow names have the first letter of each work capitalized
A DFD typically shows the minimum contents of data store. Each data store
should contain all the data elements that flow in and out.
Questionnaires should contain all the data elements that flow in and out.
Missing interfaces redundancies and like is then accounted for often through
interviews.
1. The DFD shows flow of data, not of control loops and decision are controlled
considerations do not appear on a DFD.
2. The DFD does not indicate the time factor involved in any process whether
the dataflow take place daily, weekly, monthly or yearly.
3. The sequence of events is not brought out on the DFD.
CURRENT PHYSICAL:
In Current Physical DFD proecess label include the name of people or their
positions or the names of computer systems that might provide some of the overall
system-processing label includes an identification of the technology used to process
the data. Similarly data flows and data stores are often labels with the names of
the actual physical media on which data are stored such as file folders, computer
files, business forms or computer tapes.
CURRENT LOGICAL:
The physical aspects at the system are removed as mush as possible so that
the current system is reduced to its essence to the data and the processors that
transform them regardless of actual physical form.
NEW LOGICAL:
This is exactly like a current logical model if the user were completely happy
with he user were completely happy with the functionality of the current system but
had problems with how it was implemented typically through the new logical model
will differ from current logical model while having additional functions, absolute
function removal and inefficient flows recognized.
NEW PHYSICAL:
The new physical represents only the physical implementation of the new
system.
PROCESS
1) No process can have only outputs.
2) No process can have only inputs. If an object has only inputs than it
must be a sink.
3) A process has a verb phrase label.
DATA STORE
1) Data cannot move directly from one data store to another data store, a
process must move data.
2) Data cannot move directly from an outside source to a data store, a
process, which receives, must move data from the source and place the data
into data store
3) A data store has a noun phrase label.
SOURCE OR SINK
DATA FLOW
1) A Data Flow has only one direction of flow between symbols. It may flow
in both directions between a process and a data store to show a read before an
update. The later is usually indicated however by two separate arrows since
these happen at different type.
2) A join in DFD means that exactly the same data comes from any of two or
more different processes data store or sink to a common location.
3) A data flow cannot go directly back to the same process it leads. There
must be atleast one other process that handles the data flow produce some
other data flow returns the original data into the beginning process.
4) A Data flow to a data store means update (delete or change).
5) A data Flow from a data store means retrieve or use.
A data flow has a noun phrase label more than one data flow noun phrase can
appear on a single arrow as long as all of the flows on the same arrow move
together as one package.
Add output screens here