Regression analysis: likelihood, error and entropy

2018-04-06T09:00:07Z (GMT) by Bogdan Grechuk Michael Zabarankin
In a regression with independent and identically distributed normal residuals, the log-likelihood function yields an empirical form of the L2L2-norm, whereas the normal distribution can be obtained as a solution of differential entropy maximization subject to a constraint on the L2L2-norm of a random variable. The L1L1-norm and the double exponential (Laplace) distribution are related in a similar way. These are examples of an “inter-regenerative” relationship. In fact, L2L2-norm and L1L1-norm are just particular cases of general error measures introduced by Rockafellar et al. (Finance Stoch 10(1):51–74, 2006) on a space of random variables. General error measures are not necessarily symmetric with respect to ups and downs of a random variable, which is a desired property in finance applications where gains and losses should be treated differently. This work identifies a set of all error measures, denoted by EE, and a set of all probability density functions (PDFs) that form “inter-regenerative” relationships (through log-likelihood and entropy maximization). It also shows that M-estimators, which arise in robust regression but, in general, are not error measures, form “inter-regenerative” relationships with all PDFs. In fact, the set of M-estimators, which are error measures, coincides with EE. On the other hand, M-estimators are a particular case of L-estimators that also arise in robust regression. A set of L-estimators which are error measures is identified—it contains EE and the so-called trimmed LpLp-norms.




All Rights Reserved