0% found this document useful (0 votes)
28 views5 pages

A Study On Image Annotation Techniques

This document summarizes a research paper on image annotation techniques. The paper reviews different categories of image annotation, including manual, semi-automatic, and automatic methods. Semi-automatic annotation uses contextual information from mobile devices to suggest annotations, while allowing users to improve results by providing feedback. The paper also discusses using textual information and manual tags to annotate images, but notes that manual annotation is time-consuming.

Uploaded by

IRSHAD HUSSAIN
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
28 views5 pages

A Study On Image Annotation Techniques

This document summarizes a research paper on image annotation techniques. The paper reviews different categories of image annotation, including manual, semi-automatic, and automatic methods. Semi-automatic annotation uses contextual information from mobile devices to suggest annotations, while allowing users to improve results by providing feedback. The paper also discusses using textual information and manual tags to annotate images, but notes that manual annotation is time-consuming.

Uploaded by

IRSHAD HUSSAIN
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 5

See discussions, stats, and author profiles for this publication at: https://www.researchgate.

net/publication/258650578

A Study on Image Annotation Techniques

Article  in  International Journal of Computer Applications · January 2012


DOI: 10.5120/4616-6295

CITATIONS READS

27 1,805

2 authors:

Reena Pagare Anita Shinde


MIT ART,DESIGN AND TECHNOLOGY UNIVERSITY Marathwada Mitra Mandal's College of Engineering
11 PUBLICATIONS   78 CITATIONS    7 PUBLICATIONS   48 CITATIONS   

SEE PROFILE SEE PROFILE

All content following this page was uploaded by Reena Pagare on 11 February 2016.

The user has requested enhancement of the downloaded file.


International Journal of Computer Applications (0975 – 8887)
Volume 37– No.6, January 2012

A Study on Image Annotation Techniques

Reena Pagare Anita Shinde


Assistant Professor Department Of Computer Engineering
Department Of Computer Engineering M. I. T. College Of Engineering
M. I. T. College Of Engineering

ABSTRACT suggestions. Annotations can include event, time, person,


With the development of mobile device, large amount of digital location etc [3].
images are generated every day. The requirements of effective The paper is organized as follow. Section 2 related work in
indexing and searching image are growing rapidly. Image Image Annotation. Section 3 introduces some of the commonly
annotation is an effective way for content based image used techniques in image annotations. Section 4 explains semi-
retrieval. This paper gives a review of different categories for automatic way to include annotation to image by using
image annotation. The paper also gives a brief overview of contextual information of the mobile device. Section 5
different methods for image annotation under Semi- Automatic concludes the paper.
and Automatic annotations. Manual annotation is a costly and
time consuming work, especially for mobile device. 2. RELATED WORK
Image annotation & retrieval has been a very strong area of
General Terms research nowadays. In image annotation & retrieval, text
Context, Image annotation, Retrieval. annotation acts as an important role. Efficient labeling of
photos has been an active research at the present time. In [4],
Keywords the authors spotlight on collections of personal photos and
Ontology, Semi-automatic annotation, Automatic annotation. make use of the contextual information by the associated GPS
and time metadata. The photos are annotated based on
1. INTRODUCTION collections or groups rather than individual human being. In
As the mobile communication technology and user interface [5], a background photo annotation approach on mobile device
technology are growing rapidly, people are concerned about is introduced by author. The spatial, temporal and public
various internet devices like notebook, smart phone, laptop, contexts of a photo are offered for automatic annotation. In [6],
PDA. Ubiquitous computing consists of incorporation between the inspiration of annotation is studied. It explains the different
the information space and the physical space. Because of this, ways to be taken into account while tagging their photos. In
information is accessible to people anytime & anywhere [7], the author has used the incident model in which two types
through the device, connected to internet. By this, it actually of annotation proposals are taken into account. It can be either
reduces the difficulty of working with the device & makes it for a particular user context or for a community network. In
simple & well organized. The frequently changing surrounding [8], model for photo sharing on mobile phone is explained. It
or situation in ubiquitous computing such as location is known provides annotation propositions based on web server
as context [1]. computing and user interactivity.
With the improvement of mobile device, huge amount of
digital images are produced every day. So there is need of 3. IMAGE ANNOTATION TECHNIQUES
successful indexing and searching image. Content based
method is one of the best ways for image searching. But, the 3.1 Making use of Textual Information
search via image content is still feasible field to enhance. The The huge numbers of images are available on the World Wide
main approach to search image is a keyword based way. At the Web. In [9] order to categorize and competently retrieve them ,
same time, creation of semantic metadata about photo content background information of the images such as surrounding
remains an elusive goal. The amount of annotation can greatly content and associations can be used for image annotation.
improve the usefulness of photo collections as they grow into Automatically we can obtain semantic knowledge for Web
the thousands. In addition to making the photo searchable by images. Similarly context can be assigned to web images by
the contributing user, tags enable users to discover other users‟ using page layout analysis method. At the same time, the
photos. Similarly, the usual use for annotation, personal accuracy of image retrieval will be low as it this can retrieve
organization and retrieval, is improved by the skill for users to many relevant images. There are three reasons for low
render their photos on-line to be viewed by other members of accuracy. Firstly, Web images can be used by anyone in the
the society [2]. Web pages and there is no standard exists for the relationships
between the texts and inserted images in the Web pages.
Image annotation is the professional way for content based
Secondly, Web images are fairly wide-ranging in meaning,
image retrieval. The manual annotation is high-priced and it
because they are created by different group for different
takes more time, particularly for mobile device. There are
reasons. Thirdly, the qualities of the Web images vary
successful contexts for a mobile device, like photo captured
significantly. The users require passing through the whole list
context, private context and social network context. These
of retrieved images to search the preferred ones.
contexts are combined to acquire functional semantic content
of the photo. The produced results are taken as annotation

42
International Journal of Computer Applications (0975 – 8887)
Volume 37– No.6, January 2012

3.2 Manual Annotation training examples are used as input where each training
In manual annotation users have to enter some descriptive example is described by its low level features & corresponding
keywords when the images are loaded / registered/browsed. annotation to the image. It results in template for annotating
Manual annotation of image content is considered “best case” image with set of relevant keywords for image. To bridge the
in terms of accuracy, since keywords are selected based on gap between low-level features and high-level semantics in
human determination of the semantic content of images. But at retrieval systems, user participation is required in semi
the same time, it is an effort intensive and monotonous process. automatic annotation. The user is supposed improve results by
Manual annotation can have a problem that at the retrieval, using negative as well as positive examples and revise the
users can forget the annotations they have used after a long knowledge about image classes in the semantic space.
period of time. This semi automatic [12] annotation method is combination of
efficiency of automatic annotation & accuracy of manual
3.3 Image Annotation Based On Ontology annotation. The user has to provide the feedback while
Semantic Web technologies like ontologies can be used to examining retrieval results. This method has three main parts:
annotate images with semantic descriptions. Ontology is a The query firing interface (A keyword query), the image
design of an abstract/summary. It actually defines a collection browser and the relevance feedback interface. When user
of figurative terms called concepts. Focus of ontology based submits a query, search results are returned as a ranked of
semantic image annotation is on relating the contents of an images as relevance with query. According to ranked list order,
image and tries to describe image contents as completely as images are displayed on the image browser where user can
possible. Three layer architecture [10] for image annotation is view them. After browsing images the user can give feedback
suggested. Low level features of images are selected by the through relevance feedback interface. The system returns the
bottom layer. These features are then mapped to semantically refined retrieval results based on the user‟s feedback and
significant keywords in the middle layer. These keywords are presents the results in the browser. This method is particularly
then connected to schemas and ontologies on the top layer. suitable in a dynamic database system, in which new images
are constantly being introduced continuously.
The keyword-based approach is user friendly and it can be
easily applied with satisfactory retrieval accuracy, while 3.5 Automatic Image Annotation
semantically rich ontology concentrate on the need for In this [13] automatic image annotation method, image
complete descriptions of image retrieval and advances the segmentation algorithms are used to divide the images into a
accuracy of retrieval. Ontology performs well with the number of unevenly shaped „blob‟ regions and to work on
combination of low level image features with high level textual these blobs. It uses the „global‟ features for automated image
information due to effectiveness of visual information to sort annotation. This modeling framework is based on
out the most of imprecise results. nonparametric density estimation, using the technique of
„kernel smoothing‟. User annotator has to select the word for
annotation of image with the some probability. This probability
3.4 Semi Automatic Annotation can be interpreted into probability density of image x & density
There is high semantic gap between the low level visual
of x conditional upon the assignment of annotation w.
features of the image & high level human semantic due to
which performance of traditional Content Based Information In this approach [14] a training set of images are used for
Retrieval (CBIR) systems degrades. In semi automatic image automatically annotating images. Vocabulary of blobs is used
annotation, it requires some sort of user participation in the for describing the regions in the image. In this method by using
image annotation process. the training set of images with annotated keywords, it is
required to predict the probability of deriving the label for the
It [10] explains machine learning algorithms for user supported
blobs in the image. Image can be seen as collection of blobs.
image annotation. Three layer architecture is used for image
For each, there is probability distribution called as relevance
annotation. Visual information taken from the raw image
model of image. This relevance model can be treated as
contents forms the bottom layer. These contents are them
container holding all possible blobs that exist in the image. It
mapped with the semantically rich keywords at the middle
will contain the keywords that exist in the image. With the help
layer. Top layer consists of by mapping of keywords to
of training set of images with annotated labels, the possibility
schemas (structure described in a formal language) &
of producing a tag specified the blobs in an image can be
ontologies (a formal explicit description of concepts). Machine
guessed.
learning together with user feedback helps to make use of
previously annotated images to increase the rate of annotation This method [15] uses word to word correlation because
for images from the same domain. It is consistent, cost- sometimes image features are inadequate in establishing the
effective, fast, intelligent annotation of visual data. Intelligent corresponding word annotation. To integrate the word-to-word
Image Indexing Web Service (I3WS) is used in this approach correlation, it needs to approximate the probability of
which will take a raw image repository (along with some annotating image with a set of words. This approach uses
voluntary restrictions and parameters such as schemas, language model to produce annotation words for image. This
keywords, ontologies, etc.) as an input and return its annotated model contains set of word probabilities. Probability means
version as an outcome. how probably the particular word will be used for annotation.
Advantage to this approach is that it automatically determines
In [11] this to acquire the information about the semantic
the annotation length for a given image which in turn enhances
meaning of an image called as keywords, the image is divided
precision of image retrieval.
based on contents including objects consisting of its category,
personality & its action. The resulting semantic classification
This [16] approach improves the existing annotations of images
of the image as semantic class is treated as the root of the
i.e. it will refine the conditional probability so that more
hierarchical description structure. Sequence of keywords is
accurate annotations will have higher probabilities. As effect,
used to annotate the image & selection of keyword is
the annotations with highest probabilities will be kept as the
dependent on occurrence of concept in the image. Set of

43
International Journal of Computer Applications (0975 – 8887)
Volume 37– No.6, January 2012

final annotations. For query image, an existing image


annotation method is used to obtain a set of applicant
annotations. Then, the applicant annotations are re-ranked and
only the top ones with the high probabilities are considered as
the ultimate annotations. While re-ranking it actually calculates
fixed probability of word annotating particular image.

Table 1.Comparative analysis of image annotation


Techniques

Image Advantages Disadvantages


annotation
Technique
Fig 1: Photo Context [3]

Reliable, Tedious, 4.1 Photo Captured Context


Manual Accuracy in Requires a lot of While taking a photo, the time gets recorded in camera. The
Extracting time & efforts, time is documented in the photo metadata (EXIF:
semantic Costly Exchangeable image file format). Similarly if the mobile
information at device is prepared with a GPS receiver or other positioning
several levels. system, it becomes location-aware. By using this GPS module,
the GPS coordinates of the location where the photo is taken
can also be recorded in EXIF of photo. In this way, temporal
Semi- Efficient than Requires User and spatial context from photo metadata can be achieved.
automatic manual interfaces Along with this data, there is also a field in EXIF called “artist”
annotation More refinements to which will store data about photographer. So the photographer
improve the can be recorded on the camera software.
accurate , useful feedback process
for dynamic
database
4.2 Personal Context
Being personal device, additional user context information
from mobile device can be used for annotation purpose. A
common service in mobile device is calendar that user utilize in
making their timetable. So, when a picture is taken at some
Automatic Speed (Saves Less Reliable time which is within a planned event, the picture is possibly
time) than Manual , associated to the event. Similarly, more contexts from email
More error- and contacts can be taken. One of the best semantic contents of
prone, Produces a photo is event. If the event can be acquired automatically, the
more general annotation can become easier. So it is logical to use the
scheduled event as an implication of event annotation. In
(less detailed)
Image retrieval, annotation of person in a picture is very
annotation as helpful. When it is impossible to automatically identify person
compared to in the photo, it needs to be manually annotated. The candidates
manual method for annotations can be the contacts stored in the mobile device.
Less Accurate
than automatic 4.3 Social Network Context
annotation As one of the important SNS (Social networking service)
activities is photo sharing, Social context can be used for
annotation. In social context based photo sharing system, the
friendship between users is an implied annotation of image.
4. SEMI-AUTOMATIC ANNOTATIONS This annotation needs to be added on the server side. Social
ON MOBILE PHONE USING relationship is very essential for a photo sharing system. The
ways in which friends can be added in system:
CONTEXTUAL INFORMATION
1) Adding friend by Name: The request can be sent to someone
In ubiquitous computing applications, three important aspects known to be friend by giving his user name.
of context are as location, identity and time. One more
parameter can be considered as event to deal with photos. The 2) Adding friend by Photo: While browsing photos uploaded
four important key features of photo are: time, location, person by other people, a friend request can be sent to make friendship
and event. Contextual information is used to help annotation on with owner.
mobile phone. Three kinds of context are integrated: photo 3) Importing friend from Email Contacts: Contacts of email
captured context, personal context and social network context. can be retrieved with help of email service providers. If user
wants to import friends, by inputting email and password, the
system acquires contacts from email. Then the system
evaluates these contacts with users in system, and recommends
users who are the user‟s contacts. The user will select them to
send friend requests.

44
International Journal of Computer Applications (0975 – 8887)
Volume 37– No.6, January 2012

4.4 Annotation Process system can make the annotation method more intellectual and
There is limitation of input on mobile device so the functions precise.
of annotation process are compact. The annotation algorithm
needs to be as simple as possible due to the restriction
6. REFERENCES
computing ability of mobile phone. To give annotations to an [1] Wei Liu , Xue Li, Daoli Huang “A Survey on Context
image, following are the steps: Awareness” Computer Science and Service System
(CSSS), International Conference on,29 June 2011
1) Metadata Analysis: The metadata can be acquired from the IEEE
image (EXIF), including time, GPS and artist fields. [2] B. Shevade, H. Sundaram, L. Xie. “Modeling Personal
2) Getting Personal Context: The time and GPS information is and Social Network Context for Event Annotation in
used to analyze valuable information in personal utilities such Images”. In JCDL 2007, ACM Press (2007).
as calendar, contacts and email, and the produced results are [3] Shuangrong Xia, Xiangyang Gong, Wendong Wang,Ye
presented as annotation implications; Tia “Context-Aware Image Annotation and Retrieval on
3) Tagging: Images are annotated based on these suggestions Mobile Device” 2010 IEEE.
and add other tags manually if required (such as emotion). [4] L. Cao, J. Luo, H. Kautz, T. S. Huang. “Image Annotation
within the Context of Personal Photo Collections Using
4) Uploading: The photo and the annotations are uploaded to Hierarchical Event and Scene Models”. In IEEE
the server. The server accumulates the photo and the Multimedia 2009 11(2), 208- 219.
annotations and generates multidimensional indices for the
photos. [5] W. Viana, J. B. Filho, J. Gensel, M. Villanova-Oliver, H.
Martin. PhotoMap: From location and time to context-
When a photo is selected to upload, the system reads date/time, aware photo Annotations. In Journal of Location Based
GPS and photographer (artist) from the EXIF segment of the Services 2008 2(3), 211-235
photo file. The event which is programmed at that time in the
calendar through some API (application programming [6] M. Ames, M. Naaman. “Why We Tag: Motivations for
interface) is searched. The search results are listed as event Annotation”. In proc. CHI 2007, ACM Press (2007),
annotation suggestions. 971-980.
[7] U. WESTERMANN and R. JAIN. Toward a Common
After uploading photo, the server obtains the GPS and other Event Model for Multimedia Applications. In IEEE
metadata and annotates the upload time and the owner of the Multimedia 2007 14(1), 19-29
photo. Server accumulates the photo and creates thumbnails for
the photo. Along with this, it also generates multidimensional [8] M. Davis, N. V. House, J. Towle, S. King, S. Ahern, C.
indices for the photo. These indices are used to retrieve the Burgener, Perkel, M. Finn, V.Viswanathan, M.
image. Rothenberg. “MMM2: Mobile Media Metadata for
Media Sharing”, Ext. Abstracts CHI 2005, ACM Press
(2005), 1335-1338
[9] asullah Khalid Alham, Maozhen Li1, Suhel Hammoud
and Hao Qi “Evaluating Machine Learning Techniques
for Automatic Image Annotations” 2009 IEEE
[10] O. Marques, N. Barman, Semi-Automatic Semantic
Annotation of Images Using Machine Learning
Techniques, Proc. of ISWC, pp. 550-565, 2003
[11] J. Vompras, S. Conrad, A Semi-Automated Framework
for Supporting Semantic Image Annotation, Proc. of
ISWC, pp.105-109, 2005.
[12] L. Wenyin, S. Dumais, Y. Sun, H. Zhang, M. Czerwinski
and B. Field, Semi-Automatic Image Annotation, Proc. Of
INTERACT, pp.326-333, 2001.
Fig 2: Annotation Process [13] A. Yavlinsky, E. Schofield, S. M. Rüger, Automated
1) Time Index Image Annotation using Global Features and Robust
2) User Index Nonparametric Density Estimation, Proc. of CIVR, pp.
3) Emotion Index 507-517, 2005
4) Location Index
5) Relevance Index [14] Jeon, V. Lavrenko, and R. Manmatha, Automatic Image
6) Event Index Annotation and Retrieval using Cross-Med Relevance
Models, Proc. ACM SIGIR, pp. 119-126, 2003.
5. CONCLUSION [15] R. Jin, J. Chai, L. Si, Effective Automatic Image
Manual annotation is a costly and time consuming work, Annotation via a Coherent Language Model and Active
especially for mobile device. We have discussed methods of Learning, Proc. of ACM Conference on Multimedia, pp.
automatic & semi-automatic image annotation. Semi-automatic 892-899, 2004
performs better than other annotation techniques in terms of [16] C. Wang, F. Jing, L. Zhang, H. Zhang, Content-Base
accuracy as user is participating in annotation process. The Image Annotation Refinement, Proc. of CVPR, 2007.
semi-automatic way to add annotation to image by using
contextual information of the mobile device is discussed here.
Machine learning mechanism if integrated into annotation

45

View publication stats

You might also like