This tutorial is a set of 15 hour videos by acclaimed experts in the area of machine learning and statistics from Stanford. It is suitable for Beginners as well as experts

In January 2014, Stanford University professors Trevor Hastie and Rob Tibshirani (authors of the legendary Elements of Statistical Learning textbook) taught an online course based on their newest textbook, **An Introduction to Statistical Learning with Applications in R (ISLR)**.

**As a supplement to the textbook, you may also want to watch the excellent course lecture videos** (linked below), in which Dr. Hastie and Dr. Tibshirani discuss much of the material. In case you want to browse the lecture content, I’ve also linked to the PDF slides used in the videos. For basics of Machine Learning please read through – Introduction to Machine Learning

I will highly recommend you to bookmark this page and use it as a reference whenever required.

### Chapter 1: Introduction (slides, playlist)

- Opening Remarks and Examples (18:18)
- Supervised and Unsupervised Learning (12:12)

### Chapter 2: Statistical Learning (slides, playlist)

- Statistical Learning and Regression (11:41)
- Curse of Dimensionality and Parametric Models (11:40)
- Assessing Model Accuracy and Bias-Variance Trade-off (10:04)
- Classification Problems and K-Nearest Neighbors (15:37)
- Lab: Introduction to R (14:12)

### Chapter 3: Linear Regression (slides, playlist)

## Other Preparatory Regression Tutorials

- Tutorial : Concept of Linearity in Linear Regression
- Tutorial : Linear Regression Construct
- R Tutorial : Basic 2 variable Linear Regression
- R Tutorial : Multiple Linear Regression
- R Tutorial : Residual Analysis for Regression
- R Tutorial : How to use Diagnostic Plots for Regression Models
- R Tutorial : Interpretation of R Squared and Adjusted R Squared in Regression
- R Tutorial : How to interpret F Statistic in Regression Models

- Simple Linear Regression and Confidence Intervals (13:01)
- Hypothesis Testing (8:24)
- Multiple Linear Regression and Interpreting Regression Coefficients (15:38)
- Model Selection and Qualitative Predictors (14:51)
- Interactions and Nonlinearity (14:16)
- Lab: Linear Regression (22:10)

### Chapter 4: Classification (slides, playlist)

- Introduction to Classification (10:25)
- Logistic Regression and Maximum Likelihood (9:07)
- Multivariate Logistic Regression and Confounding (9:53)
- Case-Control Sampling and Multiclass Logistic Regression (7:28)
- Linear Discriminant Analysis and Bayes Theorem (7:12)
- Univariate Linear Discriminant Analysis (7:37)
- Multivariate Linear Discriminant Analysis and ROC Curves (17:42)
- Quadratic Discriminant Analysis and Naive Bayes (10:07)
- Lab: Logistic Regression (10:14)
- Lab: Linear Discriminant Analysis (8:22)
- Lab: K-Nearest Neighbors (5:01)

### Chapter 5: Resampling Methods (slides, playlist)

- Estimating Prediction Error and Validation Set Approach (14:01)
- K-fold Cross-Validation (13:33)
- Cross-Validation: The Right and Wrong Ways (10:07)
- The Bootstrap (11:29)
- More on the Bootstrap (14:35)
- Lab: Cross-Validation (11:21)
- Lab: The Bootstrap (7:40)

### Chapter 6: Linear Model Selection and Regularization (slides, playlist)

- Linear Model Selection and Best Subset Selection (13:44)
- Forward Stepwise Selection (12:26)
- Backward Stepwise Selection (5:26)
- Estimating Test Error Using Mallow’s Cp, AIC, BIC, Adjusted R-squared (14:06)
- Estimating Test Error Using Cross-Validation (8:43)
- Shrinkage Methods and Ridge Regression (12:37)
- The Lasso (15:21)
- Tuning Parameter Selection for Ridge Regression and Lasso (5:27)
- Dimension Reduction (4:45)
- Principal Components Regression and Partial Least Squares (15:48)
- Lab: Best Subset Selection (10:36)
- Lab: Forward Stepwise Selection and Model Selection Using Validation Set (10:32)
- Lab: Model Selection Using Cross-Validation (5:32)
- Lab: Ridge Regression and Lasso (16:34)

### Chapter 7: Moving Beyond Linearity (slides, playlist)

- Polynomial Regression and Step Functions (14:59)
- Piecewise Polynomials and Splines (13:13)
- Smoothing Splines (10:10)
- Local Regression and Generalized Additive Models (10:45)
- Lab: Polynomials (21:11)
- Lab: Splines and Generalized Additive Models (12:15)

### Chapter 8: Tree-Based Methods (slides, playlist)

- Decision Trees (14:37)
- Pruning a Decision Tree (11:45)
- Classification Trees and Comparison with Linear Models (11:00)
- Bootstrap Aggregation (Bagging) and Random Forests (13:45)
- Boosting and Variable Importance (12:03)
- Lab: Decision Trees (10:13)
- Lab: Random Forests and Boosting (15:35)

### Chapter 9: Support Vector Machines (slides, playlist)

- Maximal Margin Classifier (11:35)
- Support Vector Classifier (8:04)
- Kernels and Support Vector Machines (15:04)
- Example and Comparison with Logistic Regression (14:47)
- Lab: Support Vector Machine for Classification (10:13)
- Lab: Nonlinear Support Vector Machine (7:54)

### Chapter 10: Unsupervised Learning (slides, playlist)

- Unsupervised Learning and Principal Components Analysis (12:37)
- Exploring Principal Components Analysis and Proportion of Variance Explained (17:39)
- K-means Clustering (17:17)
- Hierarchical Clustering (14:45)
- Breast Cancer Example of Hierarchical Clustering (9:24)
- Lab: Principal Components Analysis (6:28)
- Lab: K-means Clustering (6:31)
- Lab: Hierarchical Clustering (6:33)

### Interviews (playlist)

- Interview with John Chambers (10:20)
- Interview with Bradley Efron (12:08)
- Interview with Jerome Friedman (10:29)
- Interviews with statistics graduate students (7:44)

The article was originally posted at Here.

Hi Shantanu ,

You are absolutely a great and organised person.You have made life so simple by sharing all the videos and presentations from the one the best professors in the world.

I have started learning all these from here. Your website is too good and probably the most organised websites to learn data sciences.

LikeLike