statistics
Lec3 – Linear Regression: Estimation I [BDS2,ISL3,IE2&3]
In this lecture, students will learn what is linear model, how to interpret the simple regression model and know about that
the simple regression model has limitations as a general tool for empirical analysis.
Lec4 – Linear Regression: Estimation II [BDS2,ISL3,IE2&3]
In this lecture, students will learn the multiple regression models and further discuss the advantages of multiple
regressions over simple regressions. Students will also know about how to estimate the parameters in the multiple
regression models using the method of ordinary least squares and describe various statistical properties of the OLS
estimators.
Lec5 – Linear Regression: Inference [ISL3,IE4]
In this lecture, students will continues learning multiple regression analysis, and turn to the problem of testing
hypotheses about the parameters in the population regression model, which includes that testing about individual
parameters, how to test a single hypothesis involving more than one parameter, and test multiple restrictions.
Lec6 – Linear Regression: Big Data Asymptotics [IE5]
In this lecture, students will learn the asymptotic properties or large sample properties of estimators and test statistics,
and know that even without the normality assumption, t and F statistics have approximately t and F distributions, at least
in large sample sizes.
Lec7 – Linear Regression: Qualitative Information [IE7]
In this lecture, students will learn to discuss qualitative independent variables, and know about how qualitative
explanatory variables can be easily incorporated into multiple regression models. Students will also learn to discuss a
binary dependent variable.
Lec8 – Linear Regression: Heteroskedasticity [IE8]
In this lecture, students will review the consequences of heteroskedasticity for ordinary least squares estimation, and
learn the available remedies when heteroskedasticity occurs, and also know about how to test for its presence.
Lec9 – Resampling Methods [BDS1,ISL5]
In this lecture, students will know the resampling methods: cross-validation and bootstraps, and be introduced K-fold
cross-validation, nonparametric and parametric bootstraps.
Lec11 – Model Selection [BDS3,ISL6]
In this lecture, students will know why we need fitting procedures other than least squares. Best subset selection,
forward selection and backward selection will be introduced. Students will also learn two common approaches that used
to select the best model, including indirect estimation of test error, Cp, AIC, BIC, adjusted R^2, and indirect estimation
of test error, validation and cross-validation to select the best model.
Lec12 – Regularization [BDS3,ISL6]
In this lecture, students will know why we need regularization to avoid overfitting in analysis. Students will learn
regularization paths. Ridge regression and lasso regression paths will be introduced. How to use cross-validation to
select the best model will be reviewed.
Lec13 – Classification I [BDS4,ISL4]
In this lecture, students will study the approaches for predicting qualitative responses - classification. And I will explain
why linear regression is not suitable in the case of a qualitative response. Students will be introduced K nearest
neighbours, discuss the relationship between probabilities and classification, and review logistic regression.
Lec14 – Classification II [BDS4,ISL4]
In this lecture, students will study multinomial logistic regression, distributed multinomial regression and MapReduce