Untitled
Untitled
Abstract
The social network, a crucial part of our life is plagued by online impersonation
and fake accounts. According to the ‘Community Standards Enforcement
Report’ published by Facebook on March 2018, about 583 million fake
accounts were taken down just in quarter 1 of 2018 and as many as 3-4% of its
active accounts during this time were still fake. In this project, we propose a
model that could be used to classify an account as fake or genuine. This model
uses Support Vector Machine as a classification technique and can process a
large dataset of accounts at once, eliminating the need to evaluate each account
manually. The community of concern to us here is Fake Accounts and our
problem can be said to be a classification or a clustering problem. As, this is an
automatic detection method, it can be applied easily by online social networks
which has millions of profiles, whose profiles cannot be examined manually.
Introduction
Social media have radically changed the way of both news consumption and
distribution. In such a disintermediation environment, users can actively
participate in news production and thus are both producer and consumer of
news. This makes the diffusion of news more efficient via social media than
through conventional news websites. Consequently, social media have become
an important source of news for many people. According to a recent report, in
the year of 2018, roughly 68% of U.S. adults get news on social media sites
(e.g., Facebook and Twitter). However, as information is no longer verified by
journalists or experts, social media has also become a fruitful environment for
fake news (or rumors). What’s even worse, the ubiquity and easy access
characteristics of social media can dramatically accelerate the speed of fake
news spreading. Fake news has long been a critical threat to our society. It has
always been an important work for both social media companies and
government agencies to combat fake news, especially to develop ingenious and
automatic techniques that can help users differentiate between the truth and
rumors. Although a large body of research work and efforts have been focused
on fake news detection (FND) in social media, the common strategy is to
evaluate the credibility of a news report based on various features extracted
from its text content or the profile of its author. We call these methods single-
source-based methods as they focus on the properties of a single article or user.
Single-source-based methods can easily lead to detection results that are heavily
dependent on individual users and are not objective enough since it is typically
difficult to judge whether a news article is telling the truth merely based on text
words or user characteristics
Existing System
Disadvantages
In today's online social networks there have been a lot of problems like fake
profiles, online impersonation, etc. To date, no one has come up with a feasible
solution to these problems. In this project, we intend to give a framework with
which the automatic detection of fake profiles can be done so that the social life
of people become secured and by using this automatic detection technique we
can make it easier for the sites to manage the huge number of profiles, which
can't be done manually. The model provides a platform for the user where the
user can use it to detect fake profile in any social media platforms. The
proposed model can be used by users on a real-time application and to achieve
high accuracy rates in detecting fake profiles. It will hence give the users a
visualized model where the user can view the real profile and fake profile that
will be classified separately.
METHODOLOGY
DATA ACQUISITION
Admin Module
Data collection:
The first phase we do is to collect the data that we are interested in collecting
for pre-processing and to apply prediction and machine learing methods. Data
pre-processing is a data mining technique that involves transforming raw
data into an understandable format. Real world data is often incomplete,
inconsistent, and lacking certain to contain many errors. Data pre-processing is
a proven method of resolving such issues. Data pre-processing prepares raw
data for further processing. For pre-processing we have used standardization
method to pre-process the UCI dataset. This step is very important because the
quality and quantity of data that you gather will directly determine how good
your predictive model can be.
Data acquisition:
Data Preparation, where we load our data into a suitable place and prepare it for
use in our machine learning training. We’ll first put all our data together, and
then randomize the ordering
Data Set from Kaggle repository we have selected about. Supervised learning is
the method in which the machine is trained on the data which the input and
output are well labelled. The model can learn on the training data and can
process the future data to predict outcome. In our dataset we have the outcome
variable or Dependent variable i.e. Y having fake news or not
Since this model will be deployed, it is saved into a pickle file (model.pkl)
created by pickle, and this file will reflect in your project folder.
User Module
Add details
In this module user can add the details like, profile pic,nums/length
username,fullname words, nums/length fullname,name==username,description
length,external URL,private, #posts, #followers , #follows, fake etc.
View Prediction
Then user attribute details will be compared with the trained dataset values.
Then fake profile will be predicted based on attributes.
HARDWARE SPECIFICATION: (MINIMUM REQUIREMENT)
INTERNAL MEMORY CA : 2 GB
SOFTWARE SPECIFICATION:
LANGUAGE : PYTHON
BACKEND : MYSQL
Python 3.0 was released on 3 December 2008. It was a major revision of the
language that is not completely backward-compatible. Many of its major
features were backported to Python 2.6.x and 2.7.x version series. Releases of
Python 3 include the 2to3 utility, which automates (at least partially) the
translation of Python 2 code to Python 3.
Python 2.7's end-of-life date was initially set at 2015 then postponed to 2020 out
of concern that a large body of existing code could not easily be forward-ported
to Python 3.
Python is a multi-paradigm programming language. Object-oriented
programming and structured programming are fully supported, and many of its
features support functional programming and aspect-oriented programming
(including by metaprogramming and metaobjects (magic methods)). Many other
paradigms are supported via extensions, including design by contract and logic
programming.
Python uses dynamic typing and a combination of reference counting and a
cycle-detecting garbage collector for memory management. It also features
dynamic name resolution (late binding), which binds method and variable
names during program execution.
Python's design offers some support for functional programming in the Lisp
tradition. It has filter, map, and reduce functions; list comprehensions,
dictionaries, sets, and generator expressions. The standard library has two
modules (itertools and functools) that implement functional tools borrowed
from Haskell and Standard ML
Rather than having all of its functionality built into its core, Python was
designed to be highly extensible. This compact modularity has made it
particularly popular as a means of adding programmable interfaces to existing
applications. Van Rossum's vision of a small core language with a large
standard library and easily extensible interpreter stemmed from his frustrations
with ABC, which espoused the opposite approach
MYSQL
MySQL Server is a powerful database management system and the user can
create application that requires little or no programming. It supports GUI
features and an entire programming language, Phpmyadmin which can be used
to develop richer and more developed application. There are quite a few reasons,
the first being that MySQL is a feature rich program that can handle any
database related task you have. You can create places to store your data build
tools that make it easy to read and modify your database contents, and ask
questions of your data. MySQL is a relational database, a database that stores
information about related objects. In MySQL that database means a collection of
tables that hold data. It collectively stores all the other related objects such as
queries, forms and reports that are used to implement function effectively.
The MySQL database can act as a back end database for PHP as a front end,
MySQL supports the user with its powerful database management functions. A
beginner can create his/her own database very simply by some mouse clicks.
Another good reason to use MySQL as backend tool is that it is a component of
the overwhelmingly popular Open source software.
MySQL is written in C and C++. Its SQL parser is written in yacc, but it uses a
home-brewed lexical analyzer.[15] MySQL works on many system platforms,
including AIX, BSDi, FreeBSD, HP-UX, eComStation, i5/OS, IRIX, Linux,
macOS, Microsoft Windows, NetBSD, Novell NetWare, OpenBSD,
OpenSolaris, OS/2 Warp, QNX, Oracle Solaris, Symbian, SunOS, SCO
OpenServer, SCO UnixWare, Sanos and Tru64. A port of MySQL to OpenVMS
also exists.
MySQL enables data to be stored and accessed across multiple storage engines,
including InnoDB, CSV, and NDB. MySQL is also capable of replicating data
and partitioning tables for better performance and durability. MySQL users
aren't required to learn new commands; they can access their data using
standard SQL commands.
The RDBMS supports large databases with millions records and supports many
data types including signed or unsigned integers 1, 2, 3, 4, and 8 bytes long;
FLOAT; DOUBLE; CHAR; VARCHAR; BINARY; VARBINARY; TEXT;
BLOB; DATE; TIME; DATETIME; TIMESTAMP; YEAR; SET; ENUM; and
OpenGIS spatial types. Fixed- and variable-length string types are also
supported
HYPER TEXT MARKUP LANGUAGE (HTML)
Following the rigors of SGML, TBL bore HTML to the world in 1990. Since
then, many of us have it to be easy to use but sometimes quite limiting. These
limiting factors are being addressed but the World Wide Web Consortium (aka
W3c) at MIT. But HTML had to start somewhere, and its success argues that it
didn’t start out too badly.
HyperText is the method by which you move around on the web — by clicking
on special text called hyperlinks which bring you to the next page. The fact that
it is hyper just means it is not linear — i.e. you can go to any place on the
Internet whenever you want by clicking on links — there is no set order to do
things in. Markup is what HTML tags do to the text inside them. They mark it
as a certain type of text (italicised text, for example). HTML is a Language, as it
has code-words and syntax like any other language.
HTML consists of a series of short codes typed into a text-file by the site author
— these are the tags. The text is then saved as a html file, and viewed through a
browser, like Internet Explorer or Netscape Navigator. This browser reads the
file and translates the text into a visible form, hopefully rendering the page as
the author had intended. Writing your own HTML entails using tags correctly to
create your vision. You can use anything from a rudimentary text-editor to a
powerful graphical editor to create HTML pages.
The tags are what separate normal text from HTML code. You might know
them as the words between the <angle-brackets>. They allow all the cool stuff
like images and tables and stuff, just by telling your browser what to render on
the page. Different tags will perform different functions. The tags themselves
don’t appear when you view your page through a browser, but their effects do.
The simplest tags do nothing more than apply formatting to some text
Web browsers receive HTML documents from a web server or from local
storage and render the documents into multimedia web pages. HTML describes
the structure of a web page semantically and originally included cues for the
appearance of the document.
HTML elements are the building blocks of HTML pages. With HTML
constructs, images and other objects such as interactive forms may be embedded
into the rendered page. HTML provides a means to create structured documents
by denoting structural semantics for text such as headings, paragraphs, lists,
links, quotes and other items. HTML elements are delineated by tags, written
using angle brackets. Tags such as <img /> and <input /> directly introduce
content into the page. Other tags such as <p> surround and provide information
about document text and may include other tags as sub-elements. Browsers do
not display the HTML tags, but use them to interpret the content of the page.
After the HTML and HTML+ drafts expired in early 1994, the IETF created an
HTML Working Group, which in 1995 completed "HTML 2.0", the first HTML
specification intended to be treated as a standard against which future
implementations should be based.
Of course, but since making websites became more popular and needs increased
many other supporting languages have been created to allow new stuff to
happen, plus HTML is modified every few years to make way for
improvements. Cascading Stylesheets are used to control how your pages are
presented, and make pages more accessible. Basic special effects and interaction
is provided by JavaScript, which adds a lot of power to basic HTML. Most of
this advanced stuff is for later down the road, but when using all of these
technologies together, you have a lot of power at your disposal.
CSS
Cascading Style Sheets (CSS) is a style sheet language used for describing the
presentation of a document written in a markup language like HTML. CSS is a
cornerstone technology of the World Wide Web, alongside HTML and
JavaScript. CSS is designed to enable the separation of presentation and
content, including layout, colors, and fonts. This separation can improve content
accessibility, provide more flexibility and control in the specification of
presentation characteristics, enable multiple web pages to share formatting by
specifying the relevant CSS in a separate .css file, and reduce complexity and
repetition in the structural content.
Separation of formatting and content also makes it feasible to present the same
markup page in different styles for different rendering methods, such as on-
screen, in print, by voice (via speech-based browser or screen reader), and on
Braille-based tactile devices. CSS also has rules for alternate formatting if the
content is accessed on a mobile device. The name cascading comes from the
specified priority scheme to determine which style rule applies if more than one
rule matches a particular element. This cascading priority scheme is predictable.
The CSS specifications are maintained by the World Wide Web Consortium
(W3C). Internet media type (MIME type) text/css is registered for use with CSS
by RFC 2318 (March 1998). The W3C operates a free CSS validation service
for CSS documents. In addition to HTML, other markup languages support the
use of CSS including XHTML, plain XML, SVG, and XUL.
CSS has a simple syntax and uses a number of English keywords to specify the
names of various style properties. A style sheet consists of a list of rules. Each
rule or rule-set consists of one or more selectors, and a declaration block.
Before CSS, nearly all presentational attributes of HTML documents were
contained within the HTML markup. All font colors, background styles,
element alignments, borders and sizes had to be explicitly described, often
repeatedly, within the HTML. CSS lets authors move much of that information
to another file, the style sheet, resulting in considerably simpler HTML.
Stands for "Cascading Style Sheet." Cascading style sheets are used to format
the layout of Web pages. They can be used to define text styles, table sizes, and
other aspects of Web pages that previously could only be defined in a page's
HTML.
CSS helps Web developers create a uniform look across several pages of a Web
site. Instead of defining the style of each table and each block of text within a
page's HTML, commonly used styles need to be defined only once in a CSS
document. Once the style is defined in cascading style sheet, it can be used by
any page that references the CSS file. Plus, CSS makes it easy to change styles
across several pages at once. For example, a Web developer may want to
increase the default text size from 10pt to 12pt for fifty pages of a Web site. If
the pages all reference the same style sheet, the text size only needs to be
changed on the style sheet and all the pages will show the larger text.
While CSS is great for creating text styles, it is helpful for formatting other
aspects of Web page layout as well. For example, CSS can be used to define the
cell padding of table cells, the style, thickness, and color of a table's border, and
the padding around images or other objects. CSS gives Web developers more
exact control over how Web pages will look than HTML does. This is why most
Web pages today incorporate cascading style sheets.
CSS is created and maintained through a group of people within the W3C called
the CSS Working Group. The CSS Working Group creates documents called
specifications. When a specification has been discussed and officially ratified
by the W3C members, it becomes a recommendation. These ratified
specifications are called recommendations because the W3C has no control over
the actual implementation of the language. Independent companies and
organizations create that software.
JAVASCRIPT
Client-side JavaScript is the most common form of the language. The script
should be included in or referenced by an HTML document for the code to be
interpreted by the browser. It means that a web page need not be a static HTML,
but can include programs that interact with the user, control the browser, and
dynamically create HTML content. The JavaScript client-side mechanism
provides many advantages over traditional CGI server-side scripts. For
example, you might use JavaScript to check if the user has entered a valid e-
mail address in a form field. The JavaScript code is executed when the user
submits the form, and only if all the entries are valid, they would be submitted
to the Web Server. JavaScript can be used to trap user-initiated events such as
button clicks, link navigation, and other actions that the user initiates explicitly
or implicitly.
JavaScript can be implemented using JavaScript statements that are placed
within the <script>... </script> HTML tags in a web page.
You can place the <script> tags, containing your JavaScript, anywhere within
your web page, but it is normally recommended that you should keep it within
the <head> tags.
The <script> tag alerts the browser program to start interpreting all the text
between these tags as a script.
All the modern browsers come with built-in support for JavaScript. Frequently,
you may need to enable or disable this support manually. This chapter explains
the procedure of enabling and disabling JavaScript support in your browsers:
Internet Explorer, Firefox, chrome, and Opera.
Python 2.6 or higher is usually required for installation of Flask. Although Flask
and its dependencies work well with Python 3 (Python 3.3 onwards), many
Flask extensions do not support it properly. Hence, it is recommended that Flask
should be installed on Python 2.7. virtualenv is a virtual Python environment
builder. It helps a user to create multiple Python environments side-by-side.
Thereby, it can avoid compatibility issues between the different versions of the
libraries. This command needs administrator privileges. Add sudo before pip on
Linux/Mac OS. If you are on Windows, log in as Administrator. On Ubuntu
virtualenv may be installed using its package manager. The route() function of
the Flask class is a decorator, which tells the application which URL should call
the associated function. Importing flask module in the project is mandatory. An
object of Flask class is our WSGI application. Flask constructor takes the name
of current module (__name__) as argument. The rule parameter represents URL
binding with the function. The options is a list of parameters to be forwarded to
the underlying Rule object. Finally the run() method of Flask class runs the
application on the local development server.
A Flask application is started by calling the run() method. However, while the
application is under development, it should be restarted manually for each
change in the code. To avoid this inconvenience, enable debug support. The
server will then reload itself if the code changes. It will also provide a useful
debugger to track the errors if any, in the application. The Debug mode is
enabled by setting the debug property of the application object to True before
running or passing the debug parameter to the run() method.
Modern web frameworks use the routing technique to help a user remember
application URLs. It is useful to access the desired page directly without having
to navigate from the home page. The route() decorator in Flask is used to bind
URL to a function. As a result, if a user visits http://localhost:5000/hello URL,
the output of the hello_world() function will be rendered in the browser. The
add_url_rule() function of an application object is also available to bind a URL
with a function as in the above example, route() is used. It is possible to build a
URL dynamically, by adding variable parts to the rule parameter. This variable
part is marked as <variable-name>. It is passed as a keyword argument to the
function with which the rule is associated. In the following example, the rule
parameter of route() decorator contains <name> variable part attached to URL
‘/hello’. Hence, if the http://localhost:5000/hello/TutorialsPoint is entered as a
URL in the browser, ‘TutorialPoint’ will be supplied to hello() function as
argument.
An advantage of using Flask might be the fact that this framework is light, and
the risk for encountering Flask security bugs is minimal. At the same time, a
drawback might be the fact that it requires quite some effort from the part of the
programmer in order to boost the list of dependencies via plugins. A great thing
about Flask is the template engine available. The purpose of such templates is to
allow basic layout configuration for web pages with the purpose of mentioning
which element is susceptible to change. As such, you will be able to define your
template once and keep it the same all over the pages of a website. With the aid
of a template engine, you will be able to save a lot of time when setting up your
application, and even when it comes to updates or maintenance issues. Overall,
Flask is easy to learn and manage as a scalable tool. It allows any type of
approach or programming technique, as there are no restrictions included on the
app architecture or data abstraction layers. You can even run it on embedded
systems like a Raspberry Pi. Your web app can be loaded on any device,
including mobile phone, desktop pc or even a tv. Besides, it benefits from a
community that offers support and solutions suggestions to a multitude of
problems that programmers might face when using Flask in Python. The core
benefit of Flask is that the programmer controls everything, while he or she will
get a deeper understanding of how internal mechanics of frameworks function.
Werkzeug
Jinja
A framework "is a code library that makes a developer's life easier when
building reliable, scalable, and maintainable web applications" by providing
reusable code or extensions for common operations. There are a number of
frameworks for Python, including Flask, Tornado, Pyramid, and Django. Flask
is an API of Python that allows to build up web-applications. It was developed
by Armin Ronacher. Flask’s framework is more explicit than Django’s
framework and is also easier to learn because it have less base code to
implement a simple web-Application. A Web-Application Framework or Web
Framework is the collection of modules and libraries that helps the developer to
write applications without writing the low-level codes such as protocols, thread
management, etc. Flask is based on WSGI(Web Server Gateway Interface)
toolkit and Jinja2 template engine
Why Flask?
easy to use.
built in development server and debugger
integrated unit testing support
RESTful request dispatching
uses Jinja2 templating
support for secure cookies (client side sessions)
100% WSGI 1.0 compliant
Unicode based
extensively documented
Database:
A database is simply a collection of used data just like phone book. MySQL database include
such objects as tables, queries, forms, and more.
Tables:
In MySQL tables are collection of similar data. With all tables can be organized differently,
and contain mostly different information- but they should all be in the same database file. For
instance we may have a database file called video store. Containing tables named members,
tapes, reservations and so on. These tables are stored in the same database file because they
are often used together to create reports to help to fill out on screen forms.
Relational database:
MySQL is a relational database. Relational databases tools like access can help us manage
information in three important ways.
Reduce redundancy
Facilitate the sharing of information
Keep data accurate.
Fields
MySQL use key fields and indexing to help speed many database operations. We can tell
MySQL, which should be key fields, or MySQL can assign them automatically.
Queries are access objects us display, print and use our data. They can be things like field
labels that we drag around when designing reports. Or they can be pictures, or titles for
reports, or boxes containing the results of calculations.
Queries and dynasts:
Queries are request to information. When access responds with its list of data, that response
constitutes a dynaset. A dynamic set of data meeting our query criteria. Because of the way
access is designed, dynasts are updated even after we have made our query.
Forms:
Forms are on screen arrangement that make it easy to enter and read data. we can also print
the forms if we want to. We can design form our self, or let the access auto form feature.
Reports:
Reports are paper copies of dynaset. We can also print reports to disk, if we like. Access
helps us to create the reports. There are even wizards for complex printouts.
Properties:
Properties are the specification we assigned to parts of our database design. We can define
properties for fields, forms, controls and most other access objects.
- Logical design
- Physical design
Logical design reviews the present physical system, prepares input and output specifications,
makes edit security and control specifications
Physical design maps out the details of the physical system, plans, system implementation,
device a test and implementation plan.
DESIGN PROCESS
INPUT DESIGN
Input design is the process of converting the user-oriented. Input to a computer based format.
The goal of the input design is to make the data entry easier , logical and free error. Errors in
the input data are controlled by the input design. The quality of the input determines the
quality of the system output.
All the data entry screen are interactive in nature, so that the user can directly enter into data
according to the prompted messages. The user are also can directly enter into data according
to the prompted messages. The users are also provided with option of selecting an appropriate
input from a list of values. This will reduce the number of error, which are otherwise likely to
arise if they were to be entered by the user itself.
Input design is one of the most important phase of the system design. Input design is the
process where the input received in the system are planned and designed, so as to get
necessary information from the user, eliminating the information that is not required. The aim
of the input design is to ensure the maximum possible levels of accuracy and also ensures that
the input is accessible that understood by the user. The input design is the part of overall
system design, which requires very careful attention. If the data going into the system is
incorrect then the processing and output will magnify the errors.
The objectives considered during input design are:
Nature of input processing.
Flexibility and thoroughness of validation rules.
Handling of properties within the input documents.
Screen design to ensure accuracy and efficiency of the input
relationship with files.
Careful design of the input also involves attention to error handling,
controls, batching and validation procedures.
Input design features can ensure the reliability of the system and produce result from accurate
data or they can result in the production of erroneous information.
The first step is to draw a data flow diagram (DFD). The DFD was first developed by Larry
Constantine as a way of expressing system requirements in graphical form.
A DFD also known as a “bubble chart” has the purpose of clarifying system requirements and
identifying major transformations that will become programs in system design. So, it is the
starting point of the design phase that functionally decomposes the requirements
specifications down to the lowest level of detail. A DFD consists of series of bubbles join by
the data flows in the system.
The purpose of data flow diagrams is to provide a semantic bridge between users and systems
developers. The diagrams are:
• Logical representations, modeling WHAT a system does, rather than physical models
showing HOW it does it;
The goal of data flow diagramming is to have a commonly understood model of a system.
The diagrams are the basis of structured systems analysis. Data flow diagrams are supported
by other techniques of structured systems analysis such as data structure diagrams, data
dictionaries, and procedure-representing techniques such as decision tables, decision trees,
and structured English.
External Entity
An external entity is a source or destination of a data flow, which is outside the area of study.
Only those entities, which originate or receive data, are represented on a business process
diagram. The symbol used is an oval containing a meaningful and unique identifier.
Process
A process shows a transformation or manipulation of data flows within the system. The
symbol used is a rectangular box, which contains 3 descriptive elements: Firstly an
identification number appears in the upper left hand corner. This is allocated arbitrarily at the
top level and serves as a unique reference. Secondly, a location appears to the right of the
identifier and describes where in the system the process takes place.
Data Flow
A data flow shows the flow of information from its source to its destination. A data flow is
represented by a line, with arrowheads showing the direction of flow. Information always
flows to or from a process and may be written, verbal or electronic. Each data flow may be
referenced by the processes or data stores at its head and tail, or by a description of its
contents.
Data Store
A data store is a holding place for information within the system: It is represented by an open
ended narrow rectangle. Data stores may be long-term files such as sales ledgers, or may be
short-term accumulations: for example batches of documents that are waiting to be processed.
Each data store should be given a reference followed by an arbitrary number.
Resource Flow
A resource flow shows the flow of any physical material from its source to its destination.
For this reason they are sometimes referred to as physical flows. The physical material in
question should be given a meaningful name. Resource flows are usually restricted to early,
high-level diagrams and are used when a description of the physical flow of materials is
considered to be important to help the analysis.
OUTPUT DESIGN
The output form of the system is either by screen or by hard copies. Output design aims at
communicating the results of the processing of the users. The reports are generated to suit the
needs of the users .The reports have to be generated with appropriate levels. In our project
outputs are generated by asp as html pages. As its web application output is designed in a
very user-friendly this will be through screen most of the time.
CODE DESIGN
The main purpose of code design is to simplify the coding and to achieve better performance
and quality with free of errors. The coding is prepared in such a way that the internal
procedures are more meaningful validation manager is displayed for each column. The
coding of the variables is done in such a way that one other than person who developed the
packages can understand its purpose.
To reduce the server load, the project is designed in a way that most of the Validation of
fields is done as client side validation, which will be more effective.
DATABASE DESIGN
The database design involves creation of tables that are represented in physical database as
stored files. They have their own existence. Each table constitute of rows and columns where
each row can be viewed as record that consists of related information and column can be
viewed as field of data of same type. The table is also designed with some position can have a
null value.
The database design of project is designed in such a way values are kept without redundancy
and with normalized format.
DEVELOPMENT APPROACH
The importance of new system is that it is user friendly and a better interface with user’s
working on it. It can overcome the problems of manual system and the security problem.
Top down approach of software development is the incremental approach to the construction
of program structure. Modules are integrated by moving through the control hierarchy,
beginning with the main control module. Module subordinate to the main control modules is
incorporate into the structure in either a depth first or breadth first manner.
1. The main module that is overall software is divided into five modules
that are under the control of the main control module.
2. Depending on the top down approach selected subordinate stubs is
replaced one at a time with actual components.
3. Tests are conducted as each component is integrated
4. On completion of each test another stub is replaced with real time
component.
5. Regression testing may be conducted to ensure the new errors have not
been introduced.
TESTING AND IMPLEMENTATION
SYSTEM TESTING
It is the process of exercising software with the intent of finding and ultimately correcting
errors. This fundamental philosophy does not change for web applications, because web
based system and applications reside on network and inter-operate with many different
operating systems, browsers, hardware platforms and communication protocols. Thus
searching for errors is significant challenge for web applications.
Testing issues:
System testing is the state of implementation, which is aimed at ensuring that the system
works accurately and efficiently as expect before live operation, commences. It certifies that
the whole set of programs hang together System testing requires a test plan that consists of
several key activities and steps for run program, string, system and user acceptance testing.
The implementation of newly design package is important in adopting a successful new
system
Testing phase is the development phase that validates the code against the functional
specifications. Testing is a vital to the achievement of the system goals. The objective of
testing is to discover errors. To fulfill this objective a series of test step such as the unit test,
integration test, validation and system test where planned and executed.
Unit testing
Here each program is tested individually so any error apply unit is debugged. The sample
data are given for the unit testing. The unit test results are recorded for further references.
During unit testing the functions of the program unit validation and the limitations are tested.
Unit testing is testing changes made in a existing or new program this test is carried out
during the programming and each module is found to be working satisfactorily. For example
in the registration form after entering all the fields we click the submit button. When submit
button is clicked, all the data in form are validated. Only after validation entries will be added
to the database.
Unit testing comprises the set of tests performed by an individual prior to integration of the
unit into large system. The situation is illustrated in as follows
The four categories of test that a programmer will typically perform on a program unit
1. Functional test
2. Performance test
3. Stress Test
4. Structure test
Functional test involve exercising the code with nominal input values for which the
expected results are known as well as boundary values and special values.
Performance testing determines the amount of execution time spent in various parts of
unit program through put and response time and device utilization by the program.
A variation of stress testing called sensitivity testing in same situations a very small
range of data contained in a bound of valid data may cause extreme and even erroneous
processing or profound performance degradation.
Structured testing is concerned with a exercising the internal logic of a program and
traversing paths. Functional testing, stress testing performance testing are referred as “black
box” testing and structure testing is referred as “white box” testing
VALIDATION TESTING
Software validation is achieved through a serious of testes that demonstrate conformity with
requirements. Thus the proposed system under consideration has been tested by validation &
found to be working satisfactory.
OUTPUT TESTING
Asking the user about the format required by them tests the output generated by the system
under consideration .It can be done in two ways, One on screen and other on printer format.
The output format on the screen is found to be correct as the format designed n system test.
SYSTEM TESTING
In the system testing the whole system is tested for interface between each module
and program units are tested and recorded. This testing is done with sample
data. The securities, communication between interfaces are tested
1. Integrated testing
2. Acceptance testing
Integrated testing
Integrated testing is a systematic technique for constructing tests to uncover errors associated
with interface.
Objective is to take unit tested modules and build a program structure that has been dictated
by design
Acceptance testing
Acceptance testing involves planning an execution of a functional test, performance test and
stress test to verify that the implemented system satisfies the requirement.
The acceptance testing is the final stage of the user the various possibilities of the data are
entered and the results are tested.
Validation testing
Software validation is achieved through a series of test that demonstrates the conformity and
requirements. Thus the proposed system under consideration has to be tested by validation
and found to be working satisfactorily. For example in customer enters phone number field
should contain number otherwise it produces an error message similarly in all the forms the
fields are validated
Testing results
All the tests should be traceable to customer requirements the focus of testing will shift
progressively from programs Exhaustive testing is not possible To be more effective testing
should be which has probability of finding errors
QUALITY ASSURANCE
Quality assurance consists of the auditing and reporting functions of management. The goal
of quality assurance is to provide management with the data entries necessary to be informed
about the product quality thereby gaining the goal of insight and confidence that the product
quality is meeting
Quality assurance thus becomes an important component of the development process It’s
included in the industry standard (IEEE 1993) On the development process quality assurance
process is integrated into a linear development cycle through validation and verification
performed at crucial system development steps .The goals of the management is to institute
and monitor a quality assurance program with in the development process
Correctness: The extent to which the program meets the system specifications and
user objectives
Reliablility: The degree to which the system performs its intended functions overtime
Maintainability: To use with which program errors are located and corrected
Testability: The effort required a testing a program to ensure its correct performance
Portability: To ease of transporting a program from hardware configuration to another
GENERIC RISKS
Risk identification is the systematic attempt to specify threats to the project plan (estimates
the schedule resource overloading etc.). By identifying know and predictable risk the first
step is to avoiding them. When possible and controlling them when necessary there are two
types of risk.
1. Generic Risk
2. Product specific risk
Generic risks are potential threats to every software project. Only those with a clear
understanding of technology can identify product specific risk The people and the
environment that is specific to the project at a hand and to identify the product specific risk
and the project the plan and the software statement of scope are examined and answer to the
following question is developed.
What special characteristics of this product may threaten the project plan.
One method for identifying risk is to create a risk item and checklists. The checklist can be
used for risk identification and focus on some subset to know and predictable risk in the
following sub categories.
1. Product risk
2. Risk associated with overall size of software to built or modified
3. Business imparts
4. Risk associated with constraints imposed with management
5. Customer characteristics
Risk associated with sophisticated of the customer and developers ability to communicate
with the customer in a timely manner.
Technical risks
Any system developed should be secured & protected against possible hazards. Security
measures are provided to prevent unauthorized access to database at various levels. Password
protection & simple procedures to change the unauthorized access are provided to the users.
The user will have to enter the user name and password and if it is validated he can
participate in auction. Otherwise if he/she is a new user he should get registered and then he
can place an order
When he/she registered they should provide authentication through jpg files (like ration card
Xerox, voter identity card Xerox). A multi layer security architecture comprising firewalls
filtering routers encryption & digital certification must be assured in this project in real time
that order details are protected from unauthorized access.
SYSTEM IMPLEMENTATION
Implementation is the stage in the project where the theoretical design is turned into a
working system. The most crucial stage is achieving a successful new system and giving a
user confidence in that the new system will work efficiently and effectively in the
implementation stage. The stage consist of
The implementation phase is less creative than system design. A system design may be
dropped at any time prior to implementation, although it becomes more difficult when it goes
to the design phase. The final report of the implementation phase includes procedural
flowcharts, record layouts, and a workable plan for implementing the candidate system
design into a operational design.
USER TRAINING
It is designed to prepare the users for testing & converting the system. There is several ways
to trail the users they are:
1) User manual
2) Help screens
3) Training demonstrations.
1) User manual:
The summary of important functions about the system & software can be provided as a
document to the user. User training is designed to prepare the user for testing and convening
a system
The summary of important functions about the system and the software can be provided as a
document to the user
1) Documentation tools:
Document production & desktop publishing tool support nearly ever aspect of software
developers. Most software development organizations spend a substantial amount of time
developing documents, and in many cases the documentation process itself is quite
inefficient. It is not use unusual for a software development effort on documentation. For this
reason, Documentation tools provide an important opportunity to improve productivity.
2) Document restructuring:
Creating document is far too timed consuming. If the system work’s, we’ll live with what we
have. In some cases, this is the correct approach. It is not possible to recreate document for
hundreds of computer programs.
Documentation must be updated, but we have limited resources. It may not be necessary to
fully redocument an application. Rather, those portions of the system that are currently
undergoing change are fully documented.
The system is business critical and must be fully redocumented. Even in this case, an
intelligent approach is to pare documentation to an essential minimum.
SYSTEM MAINTENANCE
1. Perfective maintenance
2. Preventive maintenance
Perfective maintenance:
Preventive maintenance:
Changes made to the system to avoid future problems. Any changes can be made in the
future and our project can adopt the changes.
CONCLUSION
This paper focuses on determining the fake profile, based on various feature etc.
The analysis can be made with the help of machine learning algorithm. A
comparison can be made with the existing systems, machine learning reduces
the computational time. Hence, the treatment can begin faster. classification of
fake profile is performed. This diagnosis research work includes prediction
using algorithm like Logistic Regression, SVM with the highest accuracy. The
proposed system classify whether profile is fake or not.
FUTURE WORK
this project can be conducted with the aim of developing convolutional neural
network model to diagnose and detect fake profile. It also explored the data
augmentation technique as a pre-processing step to strengthen the classification
robustness of the CNN model. The best model, namely Inception Resnet can
achieve an average accuracy of 91%.
Data Pre-processing
Admin Fake profile
Prediction
Data
Acquisition train.csv
Data Pre-
train.csv
processing
Prediction using
train.csv
ANN
Create
Model File
User View
train.csv
Prediction
DATASET
Input Design
Output design
SAMPLE CODE
Model.py
import pandas as pd
import numpy as np
import pickle
df=pd.read_csv('train.csv')
x = cdf.iloc[:, :11]
y = cdf.iloc[:, -1]
clf=LogisticRegression()
clf.fit(x,y)
SVM = svm.LinearSVC()
SVM.fit(x, y)
print(clf.predict([[0,0.22,1,0,0,0,0,0,0,90,333]]))
print(SVM.predict([[0,0.22,1,0,0,0,0,0,0,90,333]]))
round(clf.score(x,y), 4)
round(SVM.score(x,y), 4)
file=open('my_model.pkl','wb')
pickle.dump(clf,file,protocol=2)
Main.py
import numpy as np
import pickle
file=open('my_model.pkl','rb')
clf=pickle.load(file)
#file.close()
app=Flask(__name__)
@app.route('/',methods=['GET','POST'])
def hello_world():
if request.method == 'POST':
mydict=request.form
Profile_pic=int(mydict['Profile_pic'])
userlen=int(mydict['userlen'])
fullnamewords=int(mydict['fullnamewords'])
fullnamelen=int(mydict['fullnamelen'])
uname_same=int(mydict['uname_same'])
descplen=int(mydict['descplen'])
eurl=int(mydict['eurl'])
private=int(mydict['private'])
npost=int(mydict['npost'])
nfollowers=int(mydict['nfollowers'])
nfollows=int(mydict['nfollows'])
input_feature=[Profile_pic,userlen,fullnamewords,fullnamelen,uname_sam
e,descplen,eurl,private,npost,nfollowers,nfollows]
#input_feature=[100,1,45,1,1,0]
#infprob=clf.predict_proba([input_feature])[0][1]
infprob=clf.predict_proba([input_feature])[0][1]
infprob = infprob*100
return render_template('result.html',inf=infprob)
return render_template('index.html')
if __name__ == '__main__' :
app.run(debug=False)
Template.html
<!doctype html>
<html lang="en">
<head>
<meta charset="utf-8">
<link rel="stylesheet"
href="https://maxcdn.bootstrapcdn.com/bootstrap/4.0.0/css/bootstrap.min.css"
integrity="sha384-
Gn5384xqQ1aoWXA+058RXPxPg6fy4IWvTNh0E263XmFcJlSAwiGgFAW/
dAiS6JXm" crossorigin="anonymous">
<title>Detection of Fake Profile</title>
</head>
<body><div>
<br>
</nav>
</div>
<div class="container">
<section>
<div class="container">
<div class="row">
</div>
</div>
<label>Profile pic:</label>
</select>
</div>
</div>
</div>
</div>
</div>
<label>nums/length fullname</label>
</div>
</div>
<label>name==username:</label>
</select>
</div>
</div>
<div class="row align-items-center mt-4">
<label>description length</label>
</div>
</div>
<label>external URL:</label>
<option value="1">Yes</option>
<option value="0">No</option>
</select>
</div>
</div>
<label>private:</label>
<option value="0">No</option>
</select>
</div>
</div>
<label>#posts</label>
</div>
</div>
<label>#followers</label>
</div>
</div>
<div class="row align-items-center mt-4">
<label>#follows</label>
</div>
</div>
<div class="col">
</div>
</div>
</div>
</div>
</div>
</section>
</div>
</form>
<script src="https://code.jquery.com/jquery-3.2.1.slim.min.js"
integrity="sha384-KJ3o2DKtIkvYIK3UENzmM7KCkRr/rE9/Qpg6aAZGJwFD
MVNA/GpGFF93hXpG5KkN" crossorigin="anonymous"></script>
<script
src="https://cdnjs.cloudflare.com/ajax/libs/popper.js/1.12.9/umd/popper.min.js"
integrity="sha384-ApNbgh9B+Y1QKtv3Rn7W3mgPxhU9K/ScQsAP7hUibX3
9j7fakFPskvXusvfa0b4Q" crossorigin="anonymous"></script>
<script
src="https://maxcdn.bootstrapcdn.com/bootstrap/4.0.0/js/bootstrap.min.js"
integrity="sha384-JZR6Spejh4U02d8jOt6vLEHfe/JQGiRRSQQxSfFWpi1Mqu
VdAyjUar5+76PVCmYl" crossorigin="anonymous"></script>
</body>
</html>
Rssult.html
<!doctype html>
<html lang="en">
<head>
<meta charset="utf-8">
<link rel="stylesheet"
href="https://maxcdn.bootstrapcdn.com/bootstrap/4.0.0/css/bootstrap.min.css"
integrity="sha384-
Gn5384xqQ1aoWXA+058RXPxPg6fy4IWvTNh0E263XmFcJlSAwiGgFAW/
dAiS6JXm" crossorigin="anonymous">
</head>
<body><div>
<br>
</div>
<div class="container">
{% endif %}
<script src="https://code.jquery.com/jquery-3.2.1.slim.min.js"
integrity="sha384-KJ3o2DKtIkvYIK3UENzmM7KCkRr/rE9/Qpg6aAZGJwFD
MVNA/GpGFF93hXpG5KkN" crossorigin="anonymous"></script>
<script
src="https://cdnjs.cloudflare.com/ajax/libs/popper.js/1.12.9/umd/popper.min.js"
integrity="sha384-ApNbgh9B+Y1QKtv3Rn7W3mgPxhU9K/ScQsAP7hUibX3
9j7fakFPskvXusvfa0b4Q" crossorigin="anonymous"></script>
<script
src="https://maxcdn.bootstrapcdn.com/bootstrap/4.0.0/js/bootstrap.min.js"
integrity="sha384-JZR6Spejh4U02d8jOt6vLEHfe/JQGiRRSQQxSfFWpi1Mqu
VdAyjUar5+76PVCmYl" crossorigin="anonymous"></script>
</body>
</html>
Book Reference
2) Think Python: How to Think Like a Computer Scientist, 2nd edition Allen B.
Downey (O’Reilly, 2015)
4) Real Python Course, Part 1 Real Python Team (Real Python, 2017)
WEBSITES
1) https://www.w3schools.com/python/python_reference.asp
2) http://webreference.com/programming/python/index-2.html
3) https://www.python.org/doc/
4) https://www.pythonforbeginners.com/basics/python-websites-tutorials
5) https://realpython.com/tutorials/web-dev/
6) https://www.tutorialspoint.com/python_web_development_libraries/
index.htm
7) https://pythonprogramming.net/web-development-tutorials/
8) https://www.fullstackpython.com/web-development.html
9) https://pythonspot.com/web-dev/
REFERENCE
5. muhammad imran razzak, saeeda naz and ahmad zaib deep learning for
medical image processing: overview, challenges and future
arxiv:1852.3865v2 [cs.cv] 20 july 2018