Category: Personal Thought

  • 2. As George Box famously noted:  “…the statistician knows…that in nature there never was a normal distribution, there never was a straight line, yet with normal and linear assumptions, known to be false, he can often derive results which match, to a useful approximation, those found in the real world.” (JASA, 1976, Vol. 71, 791-799)  Therefore, the normality assumption will never be exactly true when one is working with real data.
    Is linear regression valid when the outcome (dependant variable) not normally distributed?. Available from: https://www.researchgate.net/post/Is_linear_regression_valid_when_the_outcome_dependant_variable_not_normally_distributed [accessed May 17, 2015].

  • The collinearity problem in regression arises when at least one linear function of the independent variables is very nearly equal to zero. (Techni- cally, a set of vectors is collinear when a linear function is exactly equal to zero. In general discussions of the collinearity problem, the term “collinear” is often used to apply to linear functions that are only approximately zero. This convention is followed in this text.) This near-singularity may arise in several ways.

    When assessing bivariate correlations, two issues should be considered. First, correlations of even .70 (which represents “shared” variance of 50%) can impact both the explanation and estimation of the regression results. Moreover, even lower correlations can have an impact if the correlation between the two independent variables is greater than either independent variable’s correlation with the dependent measure (e.g., the situation in our earlier example of the reversal of signs).
    • The suggested cutoff for the tolerance value is. 10 (or a corresponding VIF of 10.0), which corresponds to a multiple correlation of .95 with the other independent variables. When val- ues at this level are encountered, multicollinearity problems are almost certain. However, problems are likely at much lower levels as well. For example, a VIF of 5.3 corresponds to a multiple correlation of .9 between one independent variable and all other independent vari- ables. Even a VIP of 3.0 represents a multiple correlation of .82, which would be considered high if between dependent and independent variables.
    Therefore,

    Tolerance:

    proportion of a regressor’s variance not accounted for by other regressors in the model

    low tolerance values are an indicator of multicollinearity

    Variance Inflation Factor (VIF)

    the reciprocal of the tolerance

    large VIF values are an indicator of multicollinearity

  • This tiny car can change shape and drive sideways

    แวเ

    วิืผสทืทแอมใerman researchers have developed a flexible electric car.

    Source: This tiny car can change shape and drive sideways