%0 Journal Article %A Grechuk, Bogdan %A Zabarankin, Michael %D 2018 %T Regression analysis: likelihood, error and entropy %U https://figshare.le.ac.uk/articles/journal_contribution/Regression_analysis_likelihood_error_and_entropy/10202153 %2 https://figshare.le.ac.uk/ndownloader/files/18388316 %K IR content %X In a regression with independent and identically distributed normal residuals, the log-likelihood function yields an empirical form of the L2L2-norm, whereas the normal distribution can be obtained as a solution of differential entropy maximization subject to a constraint on the L2L2-norm of a random variable. The L1L1-norm and the double exponential (Laplace) distribution are related in a similar way. These are examples of an “inter-regenerative” relationship. In fact, L2L2-norm and L1L1-norm are just particular cases of general error measures introduced by Rockafellar et al. (Finance Stoch 10(1):51–74, 2006) on a space of random variables. General error measures are not necessarily symmetric with respect to ups and downs of a random variable, which is a desired property in finance applications where gains and losses should be treated differently. This work identifies a set of all error measures, denoted by EE, and a set of all probability density functions (PDFs) that form “inter-regenerative” relationships (through log-likelihood and entropy maximization). It also shows that M-estimators, which arise in robust regression but, in general, are not error measures, form “inter-regenerative” relationships with all PDFs. In fact, the set of M-estimators, which are error measures, coincides with EE. On the other hand, M-estimators are a particular case of L-estimators that also arise in robust regression. A set of L-estimators which are error measures is identified—it contains EE and the so-called trimmed LpLp-norms. %I University of Leicester