This course offers a deep dive into the world of statistical analysis, equipping learners with cutting-edge techniques to understand and interpret data effectively. We explore a range of methodologies, from regression and classification to advanced approaches like kernel methods and support vector machines, all designed to enhance your data analysis skills.

Statistical Learning

Statistical Learning
This course is part of Introduction to Data Science Techniques Specialization

Instructor: Shahrzad (Sara) Jamshidi
Access provided by Coursera Learning Team
1,765 already enrolled
Recommended experience
Recommended experience
Intermediate level
Required Prerequisites: Python coding experience; MATH 350: Numerical Methods; MATH 474: Probability and Statistics or 475: Probability
Recommended experience
Recommended experience
Intermediate level
Required Prerequisites: Python coding experience; MATH 350: Numerical Methods; MATH 474: Probability and Statistics or 475: Probability
Skills you'll gain
- Statistical Modeling
- Statistical Machine Learning
- Supervised Learning
- Model Evaluation
- Bayesian Statistics
- Feature Engineering
- Statistical Inference
- Statistical Programming
- Unsupervised Learning
- Data Analysis
- Machine Learning
- Predictive Modeling
- Statistical Analysis
- Decision Tree Learning
- Regression Analysis
- Logistic Regression
Tools you'll learn
Details to know

Add to your LinkedIn profile
36 assignments
See how employees at top companies are mastering in-demand skills

Build your subject-matter expertise
- Learn new concepts from industry experts
- Gain a foundational understanding of a subject or tool
- Develop job-relevant skills with hands-on projects
- Earn a shareable career certificate

There are 9 modules in this course
Welcome to Statistical Learning! In this course, we will cover the topics: Statistical Learning: Terminology and Ideas, Linear Regression Methods, Linear Classification Methods, Basis Expansion Methods, Kernel Smoothing Methods, Model Assessment and Selection, Maximum Likelihood Inference, and Advanced Topics. Module 1 offers an in-depth exploration of statistical learning, beginning with the rationale behind choosing a pre-defined family of functions and optimizing the expected prediction error (EPE). It covers the essentials of statistical learning, including the loss function, the bias-variance tradeoff in model selection, and the significance of model evaluation. This module also distinguishes between supervised and unsupervised learning, discusses various types of statistical learning models and data representation, and delves into the three core elements of a statistical learning problem, providing a comprehensive introduction to this field.
What's included
8 videos5 readings4 assignments1 discussion prompt1 ungraded lab
8 videos• Total 55 minutes
- Instructor Welcome• 3 minutes
- Course Overview• 5 minutes
- Module 1 Introduction• 1 minute
- What is statistical learning?• 6 minutes
- Types of Data • 15 minutes
- Models in Statistical Learning• 7 minutes
- Model Selection • 8 minutes
- Formal Description of Statistical Learning• 11 minutes
5 readings• Total 105 minutes
- Syllabus• 10 minutes
- What is Statistical Learning Reading• 10 minutes
- Terminology and Types of Data Reading• 15 minutes
- Formal Description of Statistical Learning Reading• 60 minutes
- Module 1 Summary• 10 minutes
4 assignments• Total 38 minutes
- Module 1 Summative Assessment• 15 minutes
- What is Statistical Learning Quiz• 3 minutes
- Terminology and Types of Data Quiz• 5 minutes
- Formal Description of Statistical Learning Quiz• 15 minutes
1 discussion prompt• Total 10 minutes
- Meet and Greet Discussion• 10 minutes
1 ungraded lab• Total 60 minutes
- Coding Exercise• 60 minutes
Welcome to Module 2 of Math 569: Statistical Learning. Here, we explore what is arguably the foundational model of the field: linear regression. This simple yet highly useful model helps us better understand the statistical learning problem discussed in Module 1. In Lesson 1, we'll carefully review what linear regression aims to do, how we construct the model's parameters with a given dataset, and what kinds of statistical tests we can perform on our estimated coefficients. In Lesson 2, we’ll cover a method known as Subset Selection, which aims to improve linear regression by eliminating unimpactful independent variables. In Lesson 3, we explore introducing bias into the linear regression model with two regularization methods: Ridge Regression and LASSO. These methods utilize a hyperparameter, a key concept in this course, to limit the growth of the coefficients. This is the source of the bias and will help us understand why a biased estimator can outperform our unbiased estimator for the coefficients of linear regression in Lesson 1. Finally, Lesson 4 introduces the concept of data transformations, which allow one to address complexities within a dataset. It also provides a simple way of converting a linear model to a nonlinear model.
What's included
10 videos6 readings5 assignments6 ungraded labs
10 videos• Total 91 minutes
- Module 2 Introduction• 2 minutes
- What is Linear Regression? - Part 1• 8 minutes
- What is Linear Regression? - Part 2• 4 minutes
- Linear Regression• 11 minutes
- Linear Regression Assumptions• 10 minutes
- Statistical Tools• 21 minutes
- Subset Selection• 9 minutes
- Ridge Regression• 10 minutes
- LASSO• 9 minutes
- Data Transformation Examples and Linear Regressions • 7 minutes
6 readings• Total 290 minutes
- Module 2 Introduction Reading• 5 minutes
- Linear Regression and Least Squares Reading• 30 minutes
- Modification of Linear Regression: Subset Selection Readings• 120 minutes
- Coefficient Shrinkage for Linear Regression: Ridge Regression and LASSO Readings• 120 minutes
- Data Transformations and Linear Regression Reading• 5 minutes
- Module 2 Summary• 10 minutes
5 assignments• Total 90 minutes
- Module 2 Summative Assessment• 60 minutes
- Linear Regression and Least Squares Quiz• 10 minutes
- Modification of Linear Regression: Subset Selection Quiz• 5 minutes
- Coefficient Shrinkage for Linear Regression: Ridge Regression and LASSO Quiz• 10 minutes
- Data Transformations and Linear Regression Quiz• 5 minutes
6 ungraded labs• Total 360 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
Welcome to Module 3 of Math 569: Statistical Learning, where we delve into linear classification. In Lesson 1, we explore how linear regression, typically used for predicting continuous outcomes, can be adapted for classification tasks-predicting discrete categories. We'll cover the conversion of categorical data into a numerical format suitable for classification and introduce essential classification metrics such as accuracy, precision, and recall. In Lesson 2, we'll explore Linear Discriminant Analysis (LDA) as an alternative method for constructing linear classifications. This method introduces the notion that classification maximizes the probability of a category given a data point, a framing we will revisit later in the course. Maximizing the likelihood of classification, given some simplifying assumptions, leads to a linear model that can also reduce the dimensionality of the problem. Finally, in Lesson 3, we will cover logistic regression, which is constructed by assuming the log-likelihood odds are linear models. The outcome, similar to LDA, produces a linear decision boundary.
What's included
5 videos6 readings4 assignments6 ungraded labs
5 videos• Total 38 minutes
- Module 3 Introduction• 2 minutes
- Classification with Linear Regression• 11 minutes
- Linear Regression and Indicator Matrices• 8 minutes
- Linear Discriminant Analysis (LDA)• 10 minutes
- Logistic Regression • 8 minutes
6 readings• Total 175 minutes
- Module 3 Introduction Reading• 15 minutes
- Linear Regression of an Indicator Matrix Readings• 20 minutes
- Linear Discriminant Analysis (LDA) Readings• 45 minutes
- Logistic Regression Readings• 75 minutes
- Module 3 Summary• 10 minutes
- Insights from an Industry Leader: Learn More About Our Program• 10 minutes
4 assignments• Total 210 minutes
- Module 3 Summative Assessment• 180 minutes
- Linear Regression of an Indicator Matrix Quiz• 10 minutes
- Linear Discriminant Analysis (LDA) Quiz• 10 minutes
- Logistic Regression Quiz• 10 minutes
6 ungraded labs• Total 480 minutes
- Coding Example• 120 minutes
- Coding Exercise• 60 minutes
- Coding Example• 120 minutes
- Coding Exercise• 60 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
Welcome to Module 4 of Math 569: Statistical Learning, focusing on advanced methods in statistical modeling. This module starts with an introduction to Basis Expansion Methods, exploring how these techniques enhance linear models by incorporating non-linear relationships. We then delve into Piecewise Polynomials, discussing their utility in capturing varying trends across different segments of data. In Lesson 2, we explore Smoothing Splines, emphasizing their role in effectively balancing model fit and complexity. Lastly, Lesson 3 covers Regularization and Kernel Functions, elaborating on how these concepts contribute to constructing more complex models without significantly increasing computational complexity.
What's included
5 videos5 readings4 assignments6 ungraded labs
5 videos• Total 26 minutes
- Module 4 Introduction• 2 minutes
- What are basis expansion methods?• 3 minutes
- Piecewise Polynomials, the Method and Theory • 6 minutes
- Smoothing Splines • 6 minutes
- Regularization and Kernel Functions• 9 minutes
5 readings• Total 330 minutes
- Module 4 Introduction Reading• 20 minutes
- Piecewise Polynomials Readings• 60 minutes
- Smoothing Splines Readings• 60 minutes
- Regularization via Reproducing Kernel Hilbert Spaces Readings• 180 minutes
- Module 4 Summary• 10 minutes
4 assignments• Total 90 minutes
- Module 4 Summative Assessment• 60 minutes
- Piecewise polynomials Quiz• 10 minutes
- Smoothing Splines Quiz• 10 minutes
- Regularization via Reproducing Kernel Hilbert Spaces Quiz• 10 minutes
6 ungraded labs• Total 360 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
Welcome to Module 5 of Math 569: Statistical Learning, dedicated to advanced techniques in non-linear data modeling. In Lesson 1, we delve into Kernel Smoothers, exploring how they make predictions based on local data and their comparison to k-Nearest Neighbors (kNN) models. Lesson 2 focuses on Local Regression, particularly Local Linear Regression (LLR) and Local Polynomial Regression (LPR). We'll examine how LLR overcomes some kernel smoothing limitations and how LPR provides flexibility in capturing local data structure. The module emphasizes the adaptiveness of these techniques for complex data relationships and addresses the challenges in selecting hyperparameters and computational demands, especially for large datasets.
What's included
3 videos4 readings3 assignments4 ungraded labs
3 videos• Total 14 minutes
- Module 5 Introduction• 1 minute
- Kernel Smoothers and kNN• 7 minutes
- Local Regression • 7 minutes
4 readings• Total 140 minutes
- Module 5 Introduction Reading• 10 minutes
- Kernel Smoothers Readings• 60 minutes
- Local Regression Readings• 60 minutes
- Module 5 Summary• 10 minutes
3 assignments• Total 80 minutes
- Module 5 Summative Assessment• 60 minutes
- Kernel Smoothers Quiz• 10 minutes
- Local Regression Quiz• 10 minutes
4 ungraded labs• Total 240 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
Module 6 of Math 569: Statistical Learning delves into model evaluation and model selection via hyperparameter choice. It begins with an understanding of Bias-Variance Decomposition, highlighting the trade-off between model simplicity and accuracy. The module then explores model complexity, offering strategies for balancing this complexity with predictive performance. Building on the importance of balancing model complexity with performance, we move on to cover model selection metrics, namely: AIC, BIC, and MDL. These are information-theoretic metrics that balance error with model complexity, such as the number of parameters. Finally, the module concludes with lessons on estimating test error without a testing set, using concepts like VC Dimension, Cross-Validation, and Bootstrapping. This module is pivotal for mastering model evaluation and selection in statistical learning.
What's included
8 videos7 readings6 assignments9 ungraded labs
8 videos• Total 54 minutes
- Module 6 Introduction• 2 minutes
- Bias, Variance and Model Complexity • 10 minutes
- The Bias-Variance Decomposition• 9 minutes
- AIC and BIC • 4 minutes
- Minimum Description Length (MDL)• 7 minutes
- Vapnik-Chervonenkis (VC) Dimension • 6 minutes
- K-fold Cross Validation • 8 minutes
- Bootstrapping• 9 minutes
7 readings• Total 700 minutes
- Module 6 Introduction Readings• 15 minutes
- Bias, Variance and Model Complexity Readings• 75 minutes
- Bayesian Approach and BIC Readings• 360 minutes
- Vapnik-Chervonenkis (VC) Dimension Readings• 60 minutes
- Cross Validation Readings• 120 minutes
- Bootstrapping Readings• 60 minutes
- Module 6 Summary• 10 minutes
6 assignments• Total 340 minutes
- Module 6 Summative Assessment• 120 minutes
- Bias, Variance and Model Complexity• 10 minutes
- Bayesian Approach and BIC Quiz• 10 minutes
- Vapnik-Chervonenkis (VC) Dimension Quiz• 10 minutes
- Cross Validation Quiz• 180 minutes
- Bootstrapping Quiz• 10 minutes
9 ungraded labs• Total 540 minutes
- Coding Example• 60 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
Module 7 of Math 569: Statistical Learning introduces advanced inferential techniques. Lesson 1 focuses on Maximum Likelihood Inference, explaining how to find optimal model parameters by maximizing the likelihood function. This method is pivotal in estimating parameters for which a dataset is most likely. Lesson 2 dives into Bayesian Inference, contrasting it with frequentist approaches. It covers Bayes' Theorem, which integrates prior beliefs with new evidence to update beliefs dynamically. The module thoroughly discusses the process of Bayesian modeling, including the construction and updating of models using prior and posterior distributions. This module is crucial for understanding complex inference methods in statistical learning.
What's included
4 videos4 readings4 assignments2 ungraded labs
4 videos• Total 23 minutes
- Module 7 Introduction• 1 minute
- Maximum Likelihood Inference - Part 1• 6 minutes
- Maximum Likelihood Inference - Part 2• 7 minutes
- Bayesian Inference • 9 minutes
4 readings• Total 120 minutes
- Module 7 Introduction Reading• 5 minutes
- Maximum Likelihood Inference Reading• 45 minutes
- Bayesian Inference Readings• 60 minutes
- Module 7 Summary• 10 minutes
4 assignments• Total 260 minutes
- Module 7 Summative Assessment• 180 minutes
- Maximum Likelihood Inference Quiz- Part 1• 10 minutes
- Maximum Likelihood Inference Quiz - Part 2• 60 minutes
- Bayesian Inference Quiz• 10 minutes
2 ungraded labs• Total 120 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
Module 8 of Math 569: Statistical Learning covers diverse advanced machine learning techniques. It begins with Decision Trees, focusing on their structure and application in both classification and regression tasks. Next, it explores Support Vector Machines (SVM), detailing their function in creating optimal decision boundaries. The module then examines k-Means Clustering, an unsupervised learning method for data grouping. Finally, it concludes with Neural Networks, discussing their architecture and role in complex pattern recognition. Each lesson offers a deep dive into these techniques, showcasing their unique advantages and applications in statistical learning.
What's included
6 videos5 readings5 assignments8 ungraded labs
6 videos• Total 46 minutes
- Module 8 Introduction• 2 minutes
- Tree Models - Part 1• 7 minutes
- Tree Models - Part 2• 7 minutes
- Support Vector Machines• 10 minutes
- K-means Clustering • 6 minutes
- Neural Networks • 15 minutes
5 readings• Total 610 minutes
- Additive Models and Trees Readings• 120 minutes
- Support Vector Machines Readings• 120 minutes
- k-Means Clustering Readings• 60 minutes
- Neural Networks Readings• 300 minutes
- Module 8 Summary• 10 minutes
5 assignments• Total 100 minutes
- Module 8 Summative Assessment• 60 minutes
- Additive Models and Trees Quiz• 10 minutes
- Support Vector Machines Quiz• 10 minutes
- k-Means Clustering Quiz• 10 minutes
- Neural Networks Quiz• 10 minutes
8 ungraded labs• Total 480 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
- Coding Example• 60 minutes
- Coding Exercise• 60 minutes
This module contains the summative course assessment that has been designed to evaluate your understanding of the course material and assess your ability to apply the knowledge you have acquired throughout the course. Be sure to review the course material thoroughly before taking the assessment.
What's included
1 assignment
1 assignment• Total 180 minutes
- Course Summative Assessment• 180 minutes
Earn a career certificate
Add this credential to your LinkedIn profile, resume, or CV. Share it on social media and in your performance review.
Build toward a degree
This course is part of the following degree program(s) offered by Illinois Tech. If you are admitted and enroll, your completed coursework may count toward your degree learning and your progress can transfer with you.¹
Build toward a degree
This course is part of the following degree program(s) offered by Illinois Tech. If you are admitted and enroll, your completed coursework may count toward your degree learning and your progress can transfer with you.¹
Illinois Tech
Master of Data Science
Degree · 12-15 months
¹Successful application and enrollment are required. Eligibility requirements apply. Each institution determines the number of credits recognized by completing this content that may count towards degree requirements, considering any existing credits you may have. Click on a specific course for more information.
Instructor
Instructor ratings
We asked all learners to give feedback on our instructors based on the quality of their teaching style.

Offered by

Offered by

Illinois Tech is a top-tier, nationally ranked, private research university with programs in engineering, computer science, architecture, design, science, business, human sciences, and law. The university offers bachelor of science, master of science, professional master’s, and Ph.D. degrees—as well as certificates for in-demand STEM fields and other areas of innovation. Talented students from around the world choose to study at Illinois Tech because of the access to real-world opportunities, renowned academic programs, high value, and career prospects of graduates.
Why people choose Coursera for their career

Felipe M.

Jennifer J.

Larry W.

Chaitanya A.
Explore more from Data Science
NNortheastern University
Course

Course
NNortheastern University
Course
