0% found this document useful (0 votes)
72 views5 pages

Predicting The Missing Value by Bayesian Classification: Abstract

The document discusses predicting missing values using Bayesian classification. It begins with an introduction to data mining and extracting useful information from databases. Next, it explains why Bayesian classification is used, outlining that it allows for probabilistic learning, incremental learning, and probabilistic predictions. It then provides the Bayesian theorem formula. The rest of the document demonstrates how to apply Bayesian classification to a play tennis example, estimating probabilities of attributes given the class, and classifying an unseen sample based on attribute probabilities and class priors.

Uploaded by

Abdur Rahaman
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
72 views5 pages

Predicting The Missing Value by Bayesian Classification: Abstract

The document discusses predicting missing values using Bayesian classification. It begins with an introduction to data mining and extracting useful information from databases. Next, it explains why Bayesian classification is used, outlining that it allows for probabilistic learning, incremental learning, and probabilistic predictions. It then provides the Bayesian theorem formula. The rest of the document demonstrates how to apply Bayesian classification to a play tennis example, estimating probabilities of attributes given the class, and classifying an unseen sample based on attribute probabilities and class priors.

Uploaded by

Abdur Rahaman
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 5

Predicting The Missing Value By Bayesian Classification

Abstract:
Data Mining refers to using a variety of techniques to identify suggest of information or decision making knowledge in the database and extracting these in a way that they can put to use in areas such as decision support, predictions, forecasting and estimation. Like the healthcare industry, collects huge amounts of healthcare data which, unfortunately, are not mined to discover hidden information for effective decision making. Discovering relations that connect variables in a database is the subject of data mining. This Project consists of such kind of a System (mentioned above) where the missing values are predicted by using Bayesian Classification Technique.

Now Why Bayesian Classification?


Probabilistic learning: Calculate explicit probabilities for hypothesis, among the most
practical approaches to certain types of learning problems.

Incremental: Each training example can incrementally increase/decrease the probability


that a hypothesis is correct. Prior knowledge can be combined with observed data.

Probabilistic prediction: Predict multiple hypotheses, weighted by their probabilities. Standard: Even when Bayesian methods are computationally intractable, they can provide a
standard of optimal decision making against which other methods can be measured.

Bayes Theorem:
Given training data D, posteriori probability of a hypothesis h, P(h|D) follows the Bayes theorem

P(h | D) P(D | h)P(h) P(D)

MAP (maximum posteriori) hypothesis

h arg max P(h | D) arg max P(D | h)P(h). MAP hH hH


Practical difficulty: require initial knowledge of many probabilities, significant computational cost

Bayesian classification:
The classification problem may be formalized using posterior
probabilities:

P(C|X) = probability that the sample tuple X = <x1,,xk> belongs to the class C For example : P( class =N | outlook=sunny, windy=true,) Idea: Assign to a sample X the class label C such that P(C|X) is maximal.

Estimating a-posteriori probabilities: Bayes theorem: P(C|X) = P(X|C)P(C) / P(X) P(X) is constant for all classes P(C) = relative freq of class C samples C such that P(C|X) is maximum = C such that P(X|C)P(C) is maximum. Problem: computing P(X|C) is unfeasible!

Play-tennis example: estimating P(xi|C):


P(p)=9/14 P(n)=5/14
Outlook sunny sunny overcast rain rain rain overcast sunny sunny rain sunny overcast overcast rain Temperature Humidity Windy Class hot high false N hot high true N hot high false P mild high false P cool normal false P cool normal true N cool normal true P mild high false N cool normal false P mild normal false P mild normal true P mild high true P hot normal false P mild high true N

outlook P(sunny|p) = 2/9 P(overcast|p) = 4/9 P(rain|p) = 3/9 temperature P(hot|p) = 2/9 P(mild|p) = 4/9 P(cool|p) = 3/9 humidity P(high|p) = 3/9 P(normal|p) = 6/9 P(high|n) = 4/5 P(normal|n) = 2/5 P(hot|n) = 2/5 P(mild|n) = 2/5 P(cool|n) = 1/5 P(sunny|n) = 3/5 P(overcast|n) = 0 P(rain|n) = 2/5

windy P(true|p) = 3/9 P(false|p) = 6/9 P(true|n) = 3/5 P(false|n) = 2/5

Play-tennis example: classifying X: An unseen sample X = <rain, hot, high, false> P(X|p)P(p) = P(rain|p)P(hot|p)P(high|p)P(false|p)P(p) = 3/92/93/96/99/14 = 0.010582 P(X|n)P(n) = P(rain|n)P(hot|n)P(high|n)P(false|n)P(n) = 2/52/54/52/55/14 = 0.018286 Sample X is classified in class n (dont play).

You might also like