Module 1 - Introduction To Multimedia Databases
Module 1 - Introduction To Multimedia Databases
Give examples of multimedia applications that deals with storing, retrieving, processing and sharing of multimedia data
Eg 1. Journalism
Journalist to write article about influence of alcohol on driving Investigation involved:
Collect news articles about accidents, scientific reports, television commercials, police interviews, medical experts interviews
Illustration:
Search photo archives, stock footage companies for good photos shocking, funny, etc.
Other examples
Searching movies
Based on taste of movies already seen Based on movies a friend favor
Searching on web
Eg. searching Australian Open website (http://www.ausopen.org) Integrate conceptual terms + interesting events give info about video segments showing female American tennis players going to the net
Retrieval problems
EMPLOYEE (Name: char(20), City: Char(20), Photo: Image)
How do you select employees in Skudai? How do you select employees that wear tudung, wear glasses, fair and have a mole under the lips?
Static vs dynamic
Static: do not have time dimensions (alphanumeric data, images, graphics) Dynamic: have time dimensions (video, animation, audio)
Multimedia
Collection of media types used together At least one media types must be non-alphanumeric
Compression of text
Huffman, arithmetic coding Since storage requirements not too high, less important than multimedia data
Compression (examples)
Masking: Discard soft sound because not audible by louder sound Speech: coding of lower frequency sounds only MPEG: audio compression standards
Image compression
Exploit redundancy in image & properties of human perception
Spatial redundancy: pixels in certain area often appear similar (golden sand, blue sky) Human tolerance: error still allows effective communication
Compression of video
Compressing frames of videos: similar to image
Reduce redundancy & exploit human perception properties
Temporal redundancy: neighboring frames normally similar, remove by applying motion estimation & compression
Each image divided into fixed-sized blocks For each block in image, the most similar block in previous image is determined & pixel difference computed Together with displacement between the two blocks, this difference stored or transmitted
MPEG-1 (VHS, pixel based coding): coding of video data up to speed of 1.5 Mbits per second MPEG-2 (pixel based coding): coding of video data up to speed of 10 Mbits per second MPEG-4 (multimedia data, object based coding) : coding of video data up to speed of 40 Mbits per second, tools for decoding & representing video objects, support content-based indexing & retrieval
Structured concepts (eg semantic web, ER-like schema) can be used to describe content through concepts, their relationships to each other & MM object but
Also slow and expensive
Low-level Features
Grasp data patterns & statistics of MM object Depend strongly on medium Extraction performed automatically Eg. for text
List of keywords with frequency indicators
High-level features
Features which are meaningful to end user, such as golf course, forest How can we bridge semantic gap between low level and high level features
High level feature extraction from low level features Eg. text containing words football, referee football match text Eg. Speech to text translators (low level audio features to text) Eg. Video-Domain specific: loud sound from crowd, round object passing white line, followed by sharp whistle-goal
Similarity
Degree to which query & MM object of MIRS are similar Similarity calculated by MIRS based on metadata of MM object & query Try to estimate value of relevance of MM object to user Output is list of MM objects in descending order of similarity value
Relevance Feedback
Helps when user doesnt know exactly what he is looking for, causing problem in query formulation Interactive approach
User issue starting query, MIRS compose result set, user judge output (relevant/not), MIRS uses feedback to improve retrieval process
Starting point
query that describe info need or system provide starting point User can ask for another starting point if not satisfied Can classify object based on topics & subtopics
Quality of MIRS
Recall
r/R
r: # of relevant objects returned by system, n: # objects retrieved, R: # relevant objects in collection
Precision
r/n
Exercise
Discuss the role of DBMS in storing MM objects Discuss the role of Information Retrieval systems in storing MM objects
End of Module 1