0% found this document useful (0 votes)
51 views9 pages

Comparative Study On Semantic Search Engines

Uploaded by

Nang Nang
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
51 views9 pages

Comparative Study On Semantic Search Engines

Uploaded by

Nang Nang
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 9

See discussions, stats, and author profiles for this publication at: https://www.researchgate.

net/publication/290977643

Comparative Study on Semantic Search Engines

Article  in  International Journal of Computer Applications · December 2015


DOI: 10.5120/ijca2015907370

CITATIONS READS

6 697

3 authors, including:

Ranjna Jain Ashok Sharma


South Asian Institute of Technology and Medicine B. S. Ananagpuria Institute of Technology and management, Faridabad
2 PUBLICATIONS   6 CITATIONS    124 PUBLICATIONS   764 CITATIONS   

SEE PROFILE SEE PROFILE

Some of the authors of this publication are also working on these related projects:

PHD work of my student. View project

Design of parallel crawlers View project

All content following this page was uploaded by Ashok Sharma on 14 March 2016.

The user has requested enhancement of the downloaded file.


International Journal of Computer Applications (0975 – 8887)
Volume 131 – No.14, December2015

Comparative Study on Semantic Search Engines


Ranjna Jain Neelam Duhan, PhD A.K. Sharma, PhD
Research Scholor Assistant Professor Prof. & Dean
YMCAUST YMCAUST BSAITM
Faridabad, India Faridabad, India Faridabad, India

ABSTRACT architecture; section 4 describes the existing semantic search


Current World Wide Web also recognized as Web 2.0 is an engine‟s architecture; section 5 dictates the comparison study
immense library of interlinked documents that are transferred performed on the discussed semantic search engines and
by computers and presented to people. Search engine is finally, Section 6 concludes the paper.
considered the most important tool to discover any
information from WWW. Inspite of having lots of
2. RELATED WORK
development and novel research in current search engines The category of Search Engines includes Google, directories
techniques, they are still syntactic in nature and display search like DMOZ, yahoo etc and Meta-Search Engines such as
results on the basis of keyword matching without dogpile. They are very popular in spite of the fact that they do
understanding the meaning of query, resulting in the not provide exact results.
production of list of WebPages containing a large number of
irrelevant documents as an output. Semantic Web (Web 3.0), 2.1 Search Engine
the next version of World Wide Web is being developed with A Search Engine [3] is a program designed to search for
the aim to reduce the problem faced in Web 2.0 by information on the WWW. The search results presented in a
representing data in structured form and to discover such data list consist of web pages, images, information and other types
from Semantic Web, Semantic Search Engines (SSE) are of files. The architecture of a general search engine contains a
being developed in many domains. This paper provides a front-end process and a back-end process, as shown in Figure
survey on some of the prevalent SSEs focusing on their 1.
architecture; and presents a comparative study on the basis of
technique they follow for crawling, reasoning, indexing, Front-end
ranking etc.
Results
process
Query Search Engine Query
Interface Processor
Keywords
Web 2.0, Semantic Web, Semantic Search Engines Result
s
1. INTRODUCTION Ranking
The World Wide Web (WWW) is regarded as the largest
human information construct in history and in order to retrieve Back-end process
information from the information space, search engines are Index Files
used. The web is commonly understood to have had three
overlapping phases [2] of development and with these eras, Request
role of search engines has also changed. Under web 1.0, the WWW Crawler Indexer
purpose of search engine such as World Wide Web Worm
(WWWW) was purely on determining the size of the web and Web Pages Web Pages
content relevance was ignored. Because of the limited
resources, their indexing and hence searching were limited to
the titles and headings found in the web pages. While web 2.0 Figure 1. Architecture of General Search Engine
search engines such as google was considered as a web of In the front-end side, user submits the search query to the
documents which retrieve those documents that contain search engine interface. The query processor then parses the
keywords in that corresponding query. Due to the unstructured search request into a form that the search engine can
behaviour of information in the web page, user still had to understand, and then the search engine executes the search
mine his required information from the documents which were operation on the index files. After ranking, the search results
retrieved by the search engine on the basis of keywords. are returned to the user. In the back-end, the crawler module
Therefore, it was not successful in providing the actual (spider or robot) fetches the web pages from the Web; the
required information to the user. indexing subsystem parses those Web pages and stores them
Whereas in web 3.0 which is also named as semantic web, into the index files.
semantic search engines has web of data where data is Despite of the fact that WWW contains a lot of information
represented with triple which contain <subject- predicate- and knowledge, search engines usually serve only to deliver
object>. With this form, data is tried to be interlinked so that and present the content of documents describing the
data consumer can discover more information. It tries to knowledge. Apart from this, there exist other problems that
provide required information to the user directly so that user users are suffering from, which are discussed as follows:
does not need to explore into the displayed document as in the
case of web 2.0 search engines. The rest of the paper is  Current search engines are unable to provide direct
organized as follows: Section 2 describes Web 2.0 search answer to queries.
methods; section 3 introduces semantic web and its

4
International Journal of Computer Applications (0975 – 8887)
Volume 131 – No.14, December2015

 Current search engines process queries based on 2 where it sends user requests to several other search engines
keywords. Thus, retrieve all web pages containing and/or databases and aggregates the results into a single list
those keywords without considering the fact that an and displays them according to their source.
accurate answer is produced on the basis of user‟s
context.
 Current search engines are unable to gather complex
information.
 Current WWW contains a lot of information and
knowledge, but current search engines are unable to
retrieve complex information. For instance, user
fires a query “find 10 engineering college for
computer stream in india and the top computer
companies in their close proximity”. Current search Figure 2. Architecture of a General Meta-Search Engine
engines would not be able to yield desired results.
For the results, user has to separately fire the query Meta-Search Engines enable users to enter search criteria once
and manually merge the results. and access several search engines simultaneously. Meta-
Search engines operate on the premise that the Web is too
 Current Search Engines are handicapped by being large for any one search-engine to index it all and that more
unable to figure out the context in which a word is comprehensive search results can be obtained by combining
being used. the results from several search engines. This also may save the
user from having to use multiple search engines separately.
 Although the search engines are very helpful in
However, it is experienced by the end user that results are not
finding information on the Internet and are getting
relevant and thus, he keeps himself navigating within the
smarter with the passage of time, but they lack in
search results for a long time.
finding the meanings of the terms, expressions used
in the Web pages and the relationships between To deal with such problem, Berners-Lee, Hendler and Lassila
them. The problem comes due to the existence of [6] presented a vision of a Web in which information is given
words which have many meanings also known as well-defined meaning, better enabling computers to
polysemy and several words having same meaning understand the meaning of content and help people to provide
also known as synonymy in natura1 languages. relevant information which is called Semantic Web. The next
Thus, when a user gives a search query like “Flip- section discusses about Semantic Web.
Flop” to find the definition of “Flip-Flop” in
Computer Science domain, the most accredited 3. SEMANTIC WEB
search engine, Google, is unable to find the right Current web contains millions of unstructured web documents
document (no document is relevant among the top which are accessed using search engines such as google, bing
ten results returned). ‟ This is because Google does etc. but these search engines do not satisfy user‟s expectation
not know which Flip-Flop user is talking about; a because they display a list of documents which matches the
kind of female shoes , or a device for Electronics for terms present in the fired query. They are not concerned with
used one bit memory storage. It was possible for the fact whether they yield the user‟s required information or
Google to find the right document only if it knew not. It is because web documents are unstructured in nature
the relationship between the two terms given to it; due to which contents are analysed syntactically and thus
“Flip-Flop” and “Electronics”. makes difficult to generate the meaning of the content from it.
Therefore, to resolve this problem, Tim-Berner-Lee, inventor
2.2 Directory of WWW and director of W3C visioned about Semantic Web.
A Web Directory [4] organizes Web sites by subject, and is
usually maintained by humans instead of software. The According to Tim-Berner-Lee, Semantic Web is an extension
searcher looks at sites organized in a series of categories and of the current web in which information is given well-defined
menus. It does not display results in the form of web pages meaning, better enabling computers and people to work in co-
based on keywords. The results of directory are in the form of operation.
links that contains category and sub categories. The database The goal of semantic Web is to represent data in structured
size of directory is smaller as compared to engines' databases, format which would help machines to understand more
it is human-sited directory and not crawled by crawlers. One information on the web which supports in richer discovery and
of the famous directories, „The Open Directory‟ has been data integration from different sources via linking hereby
around since 1999, and is a human-edited directory. Also producing more exact results to the user as compared to
known as DMOZ (Directory Mozilla), the Open Directory current web search engines.
Project proposed to be the “largest on the Web”, constructed
and maintained by a vast, global community of volunteer 3.1 Architecture of Semantic Web
editors. Directory tends to work best when the user want to As part of the development of the Semantic Web vision, Tim
browse a relatively broad subject. Starting with, a directory Berners-Lee proposed a layered architecture for the Semantic
can give a good idea about the amount and type of web based Web [7] which is shown in Figure 3.The first layer consists of
information on user‟s desired topic. documents written in Unicode and their associated Uniform
Resource Identifiers (URI) and URIRef. URI is a general form
2.3 Meta-Search Engines of identifier which allow user to create and represent a
A Meta-Search engine [5] performs a search by calling on resource uniquely. It is not necessary that a URI of a resource
more than one search engine to do the actual work. The will redirect to some other location as done by URL. URLref
general architecture of Meta-Search engine is shown in Figure is another type of a string that represents aURI, and represents

5
International Journal of Computer Applications (0975 – 8887)
Volume 131 – No.14, December2015

the resource identified by that URI. It is a URI, together with <I>this is an example of current web data presentation.
the optional fragment identifies at the end separated by #.An
</BODY>
</HTML>
Rule Trust
s These tags are keywords which tell how to present data
Proof written in between them. The web browser reads the HTML
Dat

Digital Signature
a
document and uses these tags to interpret the content of the
Logic page.
Self - Dat
descripti a Whereas semantic web is an extension of the current web in
ve Ontology Vocabulary
which information is represented in structured format using
documen
RDF+rdfschema semantic language such as RDF, OWL, DAML, OIL etc.
t
XML+NS+xmlschema
For example in RDF “the sky has color blue”, will be
represented as the triple: a subject denoting the sky has color
Unicode URI+URIRef as predicate and object as value blue. Graphically, it will be
represented as shown in Figure 4.
Figure 3. Architecture of Semantic Web Has color
example of URI is http://www.w3school.org. The second Sky Blue
layer contains XML, a general purpose markup language for
documents containing structured information with XML
namespace and XML schema definitions. It makes sure that Figure 4. RDF Representation
there is a common syntax used in the semantic web. XML Each and every object is identified by its URI which helps to
schema serves for expressing schema for a particular set of resolve polysemy and synonymy problem that is often
XML documents. The third layer contains the core data encountered in current web. A lot of generic ontologies are
representation format for semantic web known as Resource available such as Dublin core, FOAF etc which are used to
Description Framework (RDF) and RDF Schema (RDFS). represent an object.
RDF represents data in the form of statement using <Subject-
Predicate-Object> that a system can process and understand. For example, The author of Book is Ranjna Jain.
RDF uses URI to identify subject-Predicate-object and to
Here Subject: Book
process these RDF statements it uses XML. RDF defines a
specific RDF/XML markup language, used to represent RDF Object: Ranjna Jain
information and for exchanging it between machines. RDF
Schema defines framework to describe classes and individuals Predicate: author
to define the vocabulary of that application. The vocabulary for the above statement is taken from Dublin
The next layer contains OWL (Web Ontology language).The core. The subject: Book can be represented by any URI such
OWL is a language derived from description logics, and offers as http://www.w3.org/Book to understand the meaning.
more constructs over RDFS. It is syntactically embedded into Here, Dublin core is used to represent the predicate: author
RDF, so like RDFS, it provides additional standardized and it will be http://purl.org/dc/elements/1.1/creator and
vocabulary. Logic and Proof layers provide the ability to finally this statement would be expressed in RDF as
perform logic on semantic statements such as inferences and
agents. Proofs are more difficult in that they must trawl many <rdf:rdf
assertions to come to conclusions. The semantic web is based
xmlns:rdf= “http://www.w3.org/1999/01/22-rdf-syntax-ns#”
on the internet. Therefore, the levels of trust in assertions and
knowledge must be determined if the source facts are to be <rdf:description rdf:about= “http://www.w3.org/book”>
believed. Digital signatures provide some trust elements, but
referrals through the “web of trust” are also valid mechanisms. <http://www.purl.org/dc/elements/1.1/creator> Ranjna Jain
A level of trust (or distrust) will need to be factored into the </http://www.purl.org/dc/elements/1.1/creator>
agents and search engines that use the semantic web. </rdf:description>
3.2 Why to represent data in a new format </rdf:rdf>
in semantic web when html is available? 4. SEMANTIC SEARCH ENGINES
Current web displays knowledge on pages using HTML which In order to access structured data, a number of semantic search
is unstructured in nature. HTML is a presentation language engines has been introduced which understand the meaning of
which displays data using tags. Web browser understands data and help in displaying more exact results as compared to
those tags and displays data accordingly but computer is not current search engines. Some of the existing prevalent
intelligent enough to understand the semantic of the content. semantic search engines have been selected for discussion in
For e.g. this section with their architectures.
<HTML> Swoogle” is a crawler-based indexing and retrieval system for
Semantic Webdocuments using RDF and OWL. It is being
<TITLE>my current page</TITLE> developed by the University of Maryland Baltimore County
<BODY> (http://pear.cs.umbc.edu/swoogle/). It extracts meta-data and
computes relations between documents. Discovered
<H1>welcome to my home page documents are alsoindexed by an information retrieval system

6
International Journal of Computer Applications (0975 – 8887)
Volume 131 – No.14, December2015

to compute the similarity among a setof documents and to search application and is restricted to retrieving
compute rank as a measure of the importance of a ontologies files with embedded RDF content on the
SemanticWeb document (SWD). internet. Apart from this, it has poor indexing of
documents and has long response time
4.1 Swoogle corresponding to fired query.
Swoogle [9] is a crawler based indexing and retrieval system
for semantic web documents (SWDs) written in RDF and 4.2 Falcon
OWL. The architecture of swoogle is discussed in Figure 5. It is a keyword based semantic search engine [10] which
Swoogle architecture can be broken into four major generates all the ranked RDF documents that include the terms
components: SWD discovery, metadata creation, data analysis in the fired query. For example user wants to know about
and interface. BSAITM, then corresponding to this query, it tries to
generates those RDF documents that contains this kind of
Interface information and in the form of snippet that exact information
IR SWD is shown so that user does not need to crawl unnecessarily to
Data
Analyze analyze other pages. It displays required information on the snippet
analysi
r r itself; therefore user does not need to explore the pages.
s Web
Server The Architecture of Falcon is described in Figure 6 and
components are described below:
Web
Service
RDF
SWD Document
Crawler
Cache SWD Agent cache
Metadat Service Web
a
Metadat
a Seeds Summarizatio
n
creation
User Interface
SWD Summar
Reader y cache
Document- level
Web Analysis
SWD Candidat
Discover e URLs Analytical Index
y Jena parser Triple Store Data
Web
Crawler
Typing analysis

Global Analysis
Textual
Description
Figure 5. Architecture of Swoogle Analysis Vocabulary Reasoning Indexing
Identification
a. SWD Discovery: At the back end, it creates a
database of SWD‟s using hybrid approach to harvest Figure 6. Architecture of Falcon Search Engine
the semantic web. It uses following mechanism to
generate URLs to find SWDs on the web: (i) seed a. RDF Crawler: An RDF crawler is setup to crawl
URLs and promising and trusted Sites (ii) URLs RDF documents. It creates queries by enumerating
from conventional search engines using meta general keywords which are sent to Google and
crawlers (iii) from swooglebot crawler that analyses swoogle to generate RDF documents. The crawler is
SWDs and generate new URI candidates. also customized to download RDF documents from
Dbpedia, Hannover, DBLP Bibliography, ping the
b. Indexing: This component indexes SWDs using its semanticweb.com
metadata and for this it captures encoding schemes
namely “RDF/XML”, N-triple, language such as b. Document level analysis: It contains jena parser,
OWL, DAML, RDFS, RDF. It records ontology which parses the cache documents collected by RDF
properties such as label, comment, version info, crawler. During this process new generated URIs
relations between two SWDs via imports, extends are queued in the seed to explore more RDF
etc. documents. Falcon index URIs by including its local
name, its associated literal values and description
c. Analysis: This component uses the created metadata about its neighboring semantic web objects in RDF
to derive analytical reports such as classification of graph and corresponding to this, maintains a virtual
SWOs and SWDB, ranking SWDs using rational document.
surfer model.
c. Global Analysis: Before indexing, vocabulary
d. Services: This interface component focuses on identification and then reasoning using class
providing data services such as search services that inclusion relation is done and then indexing is
search ontologies at the term level.But swoogle has performed.
some limitations such as; it is not a general purpose

7
International Journal of Computer Applications (0975 – 8887)
Volume 131 – No.14, December2015

d. Summarization: A query dependent snippet of characterization and by this queries are categorized
knowledge is provided to facilitate the end user to into various senses they convey.
gather its information from the snippet itself.
d. QDex Storage: It creates or maintains a file for each
e. User Interface: when a user gives a query to falcon, query which stores information about the document,
it serves a list of objects as well as types such as paragraph from which that it was extracted. After
location, organization etc. with this, user can specify that, each Qdex file is placed in a known destination
a type and focus on a particular dimension of via hash-mode operation. All this work is performed
knowledge.But Falcon has some limitations such as offline.
this engine is not interested to rank these objects
according to query. e. Query Processor: When the user fires a query from
user interface, the query is sent to the query analyzer
4.3 Hakia to generate the sense and context of the user using
Hakia[12] is a semantic search technology based search fall back algorithm and with hash mode, Qdex files
engine that presents relevant results based on concept match destination location is retrieved correctly.
rather than keyword match or popularity ranking. The f. Ranking: A pool of relevant paragraphs are ranked
Architecture of Hakia is described as below in Figure 7 and of by semantic analysis rank algorithm which is based
components described below. on advanced sentence analysis and concept match
between the query. And the best sentence for each
paragraph which will be highlighted in the snippet
to attract the user is retrieved.
WWW
But hakia has some limitation such as it has some issues such
as URL canonicalization, privacy session ID‟s, virtual
10. Results 7. Query contents and dynamic contents.

4.4 Semantic Web Search Engine


Search Engine (SWSE)[13]
1. Request 2. Crawls credible sites, Interface It is a search engine for RDF data on the web, and provides
feeds
the equivalent services a search engine currently provides for
HTML Web. The architecture is shown in Figure 8 and
component details are given below.
a. Crawler: It starts with a set of seed URIs, retrieves
Crawler Ranker Query the content of URIs, parses and writes content to
Processor
disk and recursively extracts new URIs for crawling.
9. Currently, it crawls RDF/XML syntax documents
Paragraph which are most commonly used for publishing RDF
3. WebPages
on the web.
8. Recovers
6. Queries are QDex files b. Consolidation: It provides a mean of identifying
stored equivalent entities in RDF data for e.g.; OWL
QDex System QDex defines the owl:sameas property which relates two
Storage equivalent entities; entities representing the same
real world individual but identified incongruously-
would enable the merging of information contribute
4. All possible queries
are submitted
on an entity given by heterogeneous source without
5. Meaningful queries the need for consistent URI naming of entities.
are recognized
Commercial Ontology c. Ranking: It ranks crawled data by considering URI
redirections encountered by the crawler while
performing the link based analysis.

Figure 7. Architecture of Hakia Search Engine d. Reasoning: By appending instance data(i.e.


assertional data) describing about an object, SWSE
a. Crawler: Hakia forms a collection of relevant uses scalable authoritative OWL reasoned (SAOR)
document from credible site recommended by system to infer logical consequences from a set of
librarian. It also crawl dynamic content from Blogs, facts or axioms described using classes and
news, database etc. properties and system pre-compute inference to
b. QDexing: After collecting data from different avoid the runtime expense.
segments, QDex(stands for Query Detection & e. Indexing component: It employs an inverted index
Extraction) analyzes each web page and extracts all for keyword lookups based on RDF literals (text),
the possible queries that can be asked to that page and a sparse index for lookups of structured data.
by decomposing sentences into sequences of words With a pair of keys and pointers for every entity in
resulting generating the vast number of queries. the data file, every entity in this file is associated
c. Commercial Ontology: Here, all the extracted with a particular pointer to the block in the sorted
queries are further analyzed by methods such as data file. This block contains entity snippet
morphological analysis, generalization, and containing a detailed description which is formed by

8
International Journal of Computer Applications (0975 – 8887)
Volume 131 – No.14, December2015

aggregating from many sources, description also result data for each of the hits and display them as
include inferred data which is not necessarily been an output at interface.
published but derived from the existing data through
reasoning. But Semantic Web Search engine has some limitations such
as poor ranking of documents because the Ranking process
f. Query processing and User interface: It accepts user comes before the indexing stage. Ranking technique is coming
queries, retrieves top k hits and requests the snippet independently with data indexed in dataset.

Pre-runtime

Seed Crawl Consolidate Rank Reason Index Run time


URIs

RDF Data/ Consolidated Identifier Reasoned


Redirects Data Ranks Data

Intermediary data
Query
Processing

Figure 8. Architecture of Semantic Web Search Engine

5. COMPARATIVE STUDY and relevancy of the returned pages etc. The detailed
The comparison of discussed Search Engines is performed comparison study is outlined in Table 1.
on various measures like the underlying technique, input
parameters required, working levels, complexity, quality

Table1: Comparative study of Google, Falcon, Hakia, SWSE


Google Hakia Swoogle SWSE Falcon

Evolution: Research Founded in Research project of Research Research project of


project by 2004 Ebiquity Research project of Web Software
Larry Page and is group in the Digital research group
and Sergey funded by Computer Science Enterprise at the department of
Brin, Ph.D private & Electrical Research Computer Sc. and
students at institutional Engineering Institute, Technology, Nanjing
Stanford and angel Department at the National Univ, P.R. China
working on investor University of University of
Stanford Maryland, Ireland, Galway
Digital Library Baltimore
Project County(UMBC)
Main 1. Web based Web search 1. Search semantic Keyword based Keyword based
Feature search engine generates web ontologies and search engine search engine for
operates over relevant documents for object, objects,
web pages results based 2. Searches SW operates over concepts(class and
2. Several on concept terms i.e.; URIs RDF data properties),
features such match rather 3. provides ontologies and RDF
as Query than metadata of SWDs. documents
Refinement, Keyword
Choosing match
Keywords,
everyday
essentials and
many more

9
International Journal of Computer Applications (0975 – 8887)
Volume 131 – No.14, December2015

What does HTML Web HTML Web RDF data RDF data RDF data
it Crawls? Pages Pages
Reasoning Not available Not available Rule based SAOR Class inclusion
Technique systems, Bayesian (Scalable relation
reasoning authoritative
OWL reasoner)
Indexing Inverted Qdex Swangling Inverted inverted indexing
Scheme indexing (Query Technique indexing for
Detection & literals
Extraction) sparse indexing
for structured
data
What it terms keeps a RDF triples RDF literals Local name of URIs
Indexes? collection of structured data along with its
queries (detailed associated literals
extracted description of description about
from the web entity neighbouring objects
page aggregated
from different
sources)
Ranking Page Rank Semantic Rational surfer through link Objects ranked
Technique algorithm rank model based analysis based on
algorithm combination of their
relevance to the
query and their
popularity.
Uses cosine similarity
measures b/w the
query and virtual
doc. Of objects
Knowledge Yes Yes Metadata Yes Yes
Snippet
Results document Document Entity Entity Entity

6. CONCLUSION [3] S. Brin, L. Page, The Anatomy of a Large-Scale


Web 2.0 search engines are unable to present direct answers Hypertextual Web Search Engine, Computer Networks
against the user‟s fired query because Web 2.0 contains 30 (1-7) (1998)
information which is unstructured in nature. Web 3.0 is [4] Web Directory available at:
Semantic Web using RDF format organizes information in http://en.wikipedia.org/wiki/Web_directory
more structured form which helps the semantic search engines
such as swoogle, falcon, SWSE etc. to present results in a [5] Meta-Search engine available at:
more direct way. But these semantic search engines deals with http://en.wikipedia.org/wiki/Metasearch_engine
structured data written in RDF or OWL Format only. Web 2.0 [6] Berners-Lee T., Hendler J. and Lassila O. The Semantic
also contains a huge library of interlinked documents that are Web. The Scientific American, vol. 5(1) 2001.
in semi-structured (CSV and XML) and structured form URL:http://www.scientificamerican.com/
(database). They can be used as data exchange format in
different domains- XML covers the syntactic level but lacks [7] Patel-Schneider P.F. and Fensel D. Layering the
reasoning. But, if files get converted in OWL files then they Semantic Web: Problems and Directions. In Proceedings
can be used by semantic web search engines and can expand of The Semantic Web - ISWC 2002: First International
its coverage area. This Paper has reviewed working of above Semantic Web Conference, Sardinia, Italy, vol. 2342 /
discussed Semantic Search Engines and their corresponding 2002. springer-Verlag GmbH 2002, p. 16.
techniques used in Crawling, Indexing, Ranking and Result
[8] Ora Lassila, ralph R. Swick, Resource Description
formation process.
Framework(RDF) model and syntax specification,
7. REFERENCES February 1999, www.w3.org/tr/rec-rdf-syntax-
[1] Swoogle Search Engine available at: 199990222.
http://swoogle.umbc.edu/ [9] Li Ding Tim Finin Anupam Joshi Rong Pan R. Scott Cost
[2] Comparision between phases of world Wide Web Yun Peng Pavan Reddivari Vishal Doshi Joel Sachs
available at: http://ahmedyassen.xomba.com/web-1.0- Swoogle: A Semantic Web Search and Metadata Engine.
web-2.0-web-3.0-and-web-4.0-review.html ACM Press 2004.ppp. 652-659

10
International Journal of Computer Applications (0975 – 8887)
Volume 131 – No.14, December2015

[10] Cheng, G., Ge, W., Wu, H., Qu, Y.: Searching Semantic Search Engine. Journal of Web Semantics, Vol 9 no.4
Web Objects Based on Class Hierarchies. In: Bizer, C., (2011)
Heath, T., Idehen, K., Berners-Lee, T. (eds.) LDOW
2008. CEUR-WS, vol. 369. CEUR-WS.org (2008)- [14] Services provided by falcon available at:
researchgate.net http://iws.seu.edu.cn/services/falcons/

[11] Y.Qu, G.cheng,H.Wu, W.Ge, X,Zhang, Seeking [15] T. Berners-Lee, Linked Data, Design issues for the
Knowledge with Falcon, Semantic web Challenges. World Wide Web,World WideWeb Consortium, http:
/www.w3.org/DesignIssues/LinkedData.html (2006).
[12] http://company.hakia.com/new/documents/White%20Pap
er_Semantic_Search_Technology.pdf [16] Hai Dong, Farookh Khadeer Hussain, Elizabeth Chang,A
Survey in Semantic Search Technologies, , 2008 Second
[13] Aidan Hogan, Andreas Harth , J• urgen Umbrich , Sheila IEEE International Conference on Digital Ecosystems
Kinsella , Axel Polleres , Stefan Decker, Searching and and Technologies (IEEE DEST 2008)© 2008 IEEE.
Browsing Linked Data with SWSE: the Semantic Web

IJCATM : www.ijcaonline.org 11

View publication stats

You might also like