0% found this document useful (0 votes)
17 views

Learning

Uploaded by

Priya Bhadoriya
Copyright
© © All Rights Reserved
Available Formats
Download as PDF or read online on Scribd
0% found this document useful (0 votes)
17 views

Learning

Uploaded by

Priya Bhadoriya
Copyright
© © All Rights Reserved
Available Formats
Download as PDF or read online on Scribd
You are on page 1/ 35
MACHINE LEARNING SWAPNA.C Machine Learning Learning Intelligence Def: (Learning): Acquisition of knowledge or skills through study , experience, or being taught. Def: Intelligence is the ability to learn and use concepts to solve problems. Machine Learning <> Artificial Intelligence — Def: AI is the science of making machines do things that require intelligence if done by human. (Minsky 1986). Def: Machine Learning is an area of AI concerned with development of techniques which allow machines to learn. Why Machine Learning Why Machine Learning? <> Why Artificial Intelligence? = To build machines exhibiting intelligent behaviour (i.¢., able to reason, predict, and adapt) while helping humans work, study, and entertain themselves. Recent programs in algorithm and theory. Growing flood of online data. More Computational power is available than human. Budding industry. Relevant disciplines and examples of their influence on machine learning. * Artificial intelligence Bayesian methods Computational complexity theory * Control theory Information theory * Philosophy Psychology and neurobiology Statistics Well-posed Learning Problems * Def: A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience E. * Def 2 (Hadamard 1902): A (machine learning) problem is well-posed if a solution to it exists, if that solution is unique, and if that solution depends on the data / experience but it is not sensitive to (reasonably small) changes in the data / experience. + Forexample, a computer program that learns to play checkers might improve its performance as measured by its ability to win at the class of tasks involving playing checkers games, through experience obtained by playing games against itself. Successful Applications Learning to recognize spoken words. All of the most successful speech recognition systems employ machine learning in some form. For example the SPHINX system , Learning to drive an autonomous vehicle. Machine learning methods have been used to train computer- controlled vehicles to steer correctly when driving on a variety of road types. For example, the ALVINN system. Learning to classify new astronomical structures. Machine learning methods have been applied to a variety of large databases to learn general regularities implicit in the data. For example, decision tree learning algorithms have been used by NASA to learn how to classify celestial objects from the second Palomar Observatory SkySurvey (Fayyad et al, 1995). Learning to play world-class backgammon. The most successful computer programs for playing games such as backgammon are based on machine learning algorithms. For example, the world's top computer program for backgammon, TD-GAMMO To have a well-defined learning problem Identify 3 Features: the class of tasks, the measure of performance to be improved, and the source of experience. checkers learning problem: Task T: playing checkers Performance measure P: percent of games won against opponents. Training experience E: playing practice games against itself. Some other learning problems: A handwriting recognition learning problem: * Task T: recognizing and classifying handwritten words within images * Performance measure P: percent of words correctly classified. Training experience E: a database of handwritten words with given classifications A Robot Driving Learning Problem: Task T: driving on public four-lane highways using vision sensors Performance measure P: average distance travelled before an error (as judged by human overseer) Training experience E: a sequence of images and steering commands recorded while observing a human driver. Designing a learning system 1. Choosing the training experience — Examples of best moves, games outcome ... 2. Choosing the target function — board-move, board-value, ... 3. Choosing a representation for the target function — linear function with weights (hypothesis space) 4. Choosing a learning algorithm for approximating the target function — Amethod for parameter estimation Designing A Learning System 1.Choosing the Training Experience: One key attribute is whether the training experience provides direct or indirect feedback regarding the choices made by the performance system, A second important attribute of the training experience is the degree to which the learner controls the sequence of training. A third important attribute of the training experience is how well it represents the distribution of examples over which the final system performance P must be measured. Checkers Game 2.Choosing the Target Function The next design choice is to determine exactly what type of knowledge will be learned and how this will be used by the performance program. B-->M(Priori) (Evaluation Target function) V:;B>R Let us therefore define the target value V(b) for an arbitrary board state b in B, as follows: 1. if b is a final board state that is won, then V(b) = 100 2. if b is a final board state that is lost, then V(b) =-100 3. if b is a final board state that is drawn, then V(b) =0 4. if b is a not a final state in the game, then V(b) = V(b’) where b' is the best final board state that can be achieved starting from b and playing optimally until the end of the game (assuming the opponent plays optimally, as well). 3.Choosing a Representation for the Target Function let us choose a simple representation: for any given board state, the function ¢ will be calculated as a linear combination of the following board features: xi: the number of black pieces on the board x2: the number of red pieces on the board x3: the number of black kings on the board x4: the number of red kings on the board x5: the number of black pieces threatened by red (i.e., which can be captured on red's next turn) X6: the number of red pieces threatened by black * Thus, our learning program will represent c(b) as a linear function of the form W4(b)=w0+w 1x 1+w2x2+w3x3+w4x4+w5x5+w6x6 * where wo through W6 are numerical coefficients, or weights, to be chosen by the learning algorithm. . Learned values for the weights w | through W6 will determine the relative importance of the various board features in determining the value of the board, whereas the weight w0 will provide an additive constant to the board value. Partial design of a checkers learning program: Task T: playing checkers Performance measure P; percent of games won in the world tournament. Training experience E: games played against itself Targetfunction: V:Board ->R Targetfunction representation V*(b)=w0+w 1x 1+w2x2+w3x3+w4x4+w5x5+w6x6 4.Choosing a function Approximation Algorithm In order to learn the target function f we require a set of training examples, each describing a specific board state b and the training value Vtrain(b)for b. In other words, each training example is an ordered pair of the form (b, V’,,,i,(b)). For instance, the following training example describes a board state b in which black has won the game (note x2 = 0 indicates that red has no remaining pieces) and for which the target function value Vtrain(b)is therefore +100. b=(x1=3,x2=0,x3=1x4=0,x5=0x6=0) <>. Estimating training values, Adjusting the Weights 4.1Estimating Training Values ESTIMATING TRAINING VALUES: In ex. Training information only available is about game was won or not. An approach is made to assign the training value of Vtrain(b) for intermediate board state b to be V*(successor(b)). Where v(b) the learners current approximation to V.(Successor(b) denotes the next board state of b. V,,,i.(b)c¢ c(~successor(b)) Each training example is an ordered pair of the form k for estimating training values< b, Vtrain (b)> Recall that according to our formulation of the learning problem, the only training information available to our learner is whether the game was eventually won or lost. Rule for estimating training values. Vtrain(b)V*(Successor(b)) 124) ADJUSTING THE WEIGHTS Al that remains is to specify the learning algorithm for choosing the weights w; t, best fit the set of training examples {(b, Viain(b))}. As a first step we must define what we mean by the best fit to the training data, One common approach is to define the best hypothesis, or set of weights, as that which minimizes the squared error E between the training values and the values predicted by the hypothesis V. B= Plt) — VO {b,Vria(b))€ training examples Adjusting The Weights * Several algorithms are known for finding weights of a linear function that minimize E defined in this way. * In our case, we require an algorithm that will incrementally refine the weights as new training examples become available and that will be robust to errors in these estimated training values. One such algorithm is called the Least Mean Squares, or LMS training rule. * For each observed training example it adjusts the weights a small amount in the direction that reduces the error on this training example. LMS weight update rule. For each training example (b, Vi-aix(b)) © Use the current weights to calculate V(b) © For cach weight w,, update it as wy © wy, +9 (Vraia(b) — V(b) 35 Here 7 is a small constant (e.g., 0.1) that moderates the size of the weight update. To get an intuitive understanding for why this weight update rule works, notice that when the error (Vjrin(b) — V(b) is zero, no weights are changed. When (Virain(b) — V(b) is positive (Le., when (0) is too low), then each weight is increased in proportion to the value of its corresponding feature. This will raise the value of (6), reducing the error. Notice that if the value of some feature x, is zero, then its weight is not altered regardless of the error, so that the only 4.3Final Design * The Final design of our checkers learning system can be naturally described by four distinct program modules that represent the central components in many learning systems. * Performance System : It is the module that must solve the given performance task, in this case playing checkers, by using the learned target functions(s). ‘Training examples bp Way (B;) % Kye Wig Bo) > ol Solution tne (game history) FIGURE 1.1 Final design of the checkers learning program. The Critic takes as input the history or trace of the game and produces as output a set of training examples of the target function [he Generalizer takes as input the training examples and produces an output hypothesis that is its estimate of the target function. The Experiment Generator takes as input the current hypothesis (currently learned function) and outputs a new problem (i.e., initial board state) for the Performance System to explore. Its role is to pick new practice problems that will maximize the learning rate of the overall system. In our example, the Experiment Generator follows a very simple strategy: It always proposes the same initial game board to begin a new game. Determine Type of Training Experience Games against “Fable of correct mowes, Games against wel Determine ‘Target Function Determine Representation of Learned Function Polynomial = ‘Linear function GURE 1.2 Summary of choices in designing the checkers learning program. Perspective & Issues in Machine Learning Perspective: * It involves searching a very large space of possible hypothesis to determine the one that best fits the observed data. Issues: * Which algorithm performs best for which types of problems & representation? * How much training data is sufficient? * Can prior knowledge be helpful even when it is only approximately correct? * The best strategy for choosing a useful next training experience. * What specific function should the system attempt to learn? + How can learner automatically alter it’s representation to improve it’s ability to represent.and learn the target function? ML Study PAC Learning 2014.09.11 Sanghyuk Chun Stato. tation and Search “Teaveting Satesman aut ygags Dye, oa al of cnn on tn tte ha rout a te pana, ‘ings“Foo mips ore citfetent tea ana reat rer the Cate, Atte Doping Sma atl te ape ror Ore laftnaat Getuoan, ana since Euan tye onan hee ‘Sica araciotay reine ns Se by ne uateacs me cine aroma he ip Sweerany mast ty niea orl sree ee” pirgistaass tay areal Saree tes ce care bows cet coun we near a pees ate SLIDELL tae aekasany fark man nce are Dr. Varun Kumar Decision Tree R. Akerkar TMRF, Kolhapur, India Air ANALYTICAL LEARNING SWAPNA.C INDUCTIVE BIAS Swapna.C Asst.Prof. IT Dept. Sridevi Women’s Engineering College 2-Comnider the image segment shown Below. (a}4¢ We (0,2) and compote 04. Jn Ore alstanens Betaroan Ba 9) By 9) = wt ‘0510. = mmaat xp = x0 vO =ya = amc, 3 3 ~~ PRESENTED BY, JFRIEDA, R.MANI MEGALAI, MAMUTHULAKSHM, MPhil(CSE), MSUNIVERSITY, TIRUNELVELL

You might also like