Probabilisic interpretaion

1. Guide

    When faced with a regression problem, why might linear regression, and specifically why might the least-squares cost function J, be a reasonable choice? In this section, we will give a set of probabilistic assumptions, under which least-squares regression is derived as a very natural algorithm.

 

2. Let us assume that the target variables and the inputs are related via the equation:

                               y(i) = θT x(i) + ε(i),

    where e(i) is an error term that captures either unmodeled effects (such as if there are some features very pertinent to predicting housing price, but we’d left out of the regression), or random noise.

    We assume ε(i) distributed IID (independently and identically distributed), and ε(i) ∼ N(0, σ2), the density of ε(i) is given by

                        

    We know: ε(i) = y(i) - θT x(i), this implies that:

              

    The notation “p(y(i)|x(i); θ)” indicates that this is the distribution of y(i) given x(i) and parameterized by θ. Note that we should not condition on θ (“p(y(i)|x(i), θ)”), since θ is not a random variable. We can also write the distribution of y(i) as as y(i) | x(i); θ ∼ N(θT x(i), σ2).

    We consider X(X is a matrix contains all the x(i)), ~y(y is a vector contains all the y(i)), we can get the likelihood function(now θ is a random variable):

                                      

    Notice, ε(i)  is indepedence, so y(i) | x(i) ; θ is indepedence, and y(i) | x(i); θ ∼ N(θT x(i), σ2)

                                

    The principal of maximum likelihood says that we should should choose θ so as to make the data as high probability as possible. I.e., we should choose θ to maximize L(θ).

    In particular, the derivations will be a bit simpler if we instead maximize the log likelihood l(θ):

                             

    Hence, maximizing l(θ) gives the same answer as minimizing

              

    which we recognize to be J(θ), our original least-squares cost function.

 

3. Summarize:

   Under the previous probabilistic assumptions on the data, least-squares regression corresponds to finding the maximum likelihood estimate of θ. This is thus one set of assumptions under which least-squares regression can be justified as a very natural method that’s just doing maximum likelihood estimation. (Note however that the probabilistic assumptions are by no means necessary for least-squares to be a perfectly good and rational procedure, and there may—and indeed there are—other natural assumptions that can also be used to justify it.)

    Note also that, in our previous discussion, our final choice of θ did not depend on what was σ2, and indeed we’d have arrived at the same result even if σ2 were unknown. We will use this fact again later, when we talk about the exponential family and generalized linear models.

 

    

    

posted on 2013-04-13 09:44  BigPalm  阅读(186)  评论(0编辑  收藏  举报

导航