Recently, in regression analysis, the sign of correlation coefficient is opposite to that of regression equation coefficient. After research, it is confirmed that it is a multicollinearity problem and the solution is explored.
Here, the related knowledge of multicollinearity is sorted out as follows.
It is possible that the two explanatory variables are highly correlated in theory, but the observed values may not be highly correlated, and vice versa. So multicollinearity is essentially a data problem.
There are several reasons for multicollinearity
1. All explanatory variables share the same time trend;
2. One explanatory variable is the lag of the other, and they tend to follow the same trend;
3. Because the basis of data collection is not wide enough, some explanatory variables may change together;
4. There is a linear relationship between some explanatory variables;
distinguish:
1. It is found that the sign of coefficient estimation is not correct;
2. Some important explanatory variables t value is low, but r square is not low
3. When an unimportant explanatory variable was deleted, the regression results changed significantly;
Inspection;
1. In correlation analysis, the correlation coefficient higher than 0.8 indicates the existence of multicollinearity, but the low correlation coefficient does not indicate the absence of multicollinearity;
2. Vif test;
3. Conditional coefficient test;
resolvent:
1. Increase data;
2. Some constraints are imposed on the model;
3. Delete one or more collinear variables;
4. Deform the model properly;
5. Principal component regression
The principle of dealing with multicollinearity is as follows
1. Multicollinearity is universal, and no measures can be taken for minor multicollinearity problems;
2. Serious multicollinearity problems can be found by experience or regression analysis. Such as the sign of influence coefficient, the important explanatory variable t value is very low. Necessary measures should be taken according to different situations.
3. If the model is only used for prediction, as long as the fitting degree is good, it can not deal with the multicollinearity problem. When the multicollinearity model is used for prediction, it often does not affect the prediction results;
Above is excerpt “econometrics intermediate course” pan Shengchu chief editor
Read More:
- Python normality test: test whether the data obey the normal distribution
- import org.junit.Test And @ test error reporting — solutions
- RuntimeError: CUDA error: out of memory solution (valid for pro-test)
- Solution of server connection reused (test environment)
- The usage of Java JUnit unit test and the solution of initialization error
- Attributeerror: ‘STR’ object has no attribute ‘decode’ solution: the pro test is successful.
- Monkey stress test (4)-adb opendir failed, Permission denied solution
- Clion configuration MingGW report error test cmake run with errors solution
- Solution to shell script invocation error during Xcode real machine test
- Google browser chrome installation failed, error code: 0xa0430721 solution (★ pro test is feasible 〉)
- Solution to the problem of spring boot running test class error creating bean with name ‘serverendpoint exporter’ defined
- Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.22.1:test (default-test)
- After the new video card rtx3060 arrives, configure tensorflow and run “TF. Test. Is”_ gpu_ The solution of “available ()” output false
- Maven skip unit test- maven.test.skip And skipstests
- Troubleshooting of Jenkins upgrade to 2.42 JUnit test report
- Failed to retrieve platformtransactionmanager for @ transactional test: [defaulttest]
- Maven skip unit test
- Go test print no output
- Causes and solutions of nm test o file format not recognized
- [Solved] Unit test automatically injects error reporting nullpointer