# Comparing residuals between OLS and non-OLS regressions

Suppose you want to estimate a linear model: ($n$ observations of the response, and $p+1$ predictors)

One way to do this is through the OLS solution, i.e. choose the coefficients so that the sum of square errors is minimum:

Alternatively, you could use another loss function, like the sum of the absolute deviations, so that:

Suppose you have found the parameters for the two models, and want to choose the model with the smallest value of the loss function. How can you compare the minimum values attained by the loss functions in general? (i.e. not just this specific case – we could also try other $L_p$ based loss functions) There seems to be a difference in the scale of the functions – one deals with squares while the other does not.