Basic Course Information
Welcome to our Machine Learning/Deep Learning course in school year 2021-2022. In the Fall semester, we cover classical machine learning, and in the Spring semester, we cover deep learning (primarily based upon neural networks). The current course on data mining was originally developed in 2016, and this is its 5th edition.
This course covers basics for statistical learning, with an emphasis on its application to genomic data. As a first course in statistical learning, we will largely follow the flow of the first two books (both of which are free for download! The 2nd book ISL is a baby-version of the 1st book ESL):
1. The Elements of Statistical Learning. (ESL) Authors: Hastie T., Tibshirani R. and Friedman J.
2. An Introduction of Statistical Learning with Applications in R. (ISL) Authors: James, G., Witten, D., Hastie, T., Tibshirani, R (ps: the 2nd edition of the book is out and free to download).
A few topics covered in 1 and 2 appear to be obsolete (e.g. boosting algorithms and support vector machines), so we will supplement with more up-to-date materials from other books or original papers.
A few other references are helpful to have, including
3. Deep Learning by Ian Goodfellow (the first section of the book talks about the math background information as well as classical machine learning)
4. Deep Medicine by Eric Topol (This is a non-technical book, talking about numerous cool applications of AI to medicine).
5. Statistical Learning with Sparsity (This is a technical book, covering the theoretical aspects of LASSO and its further extensions)
Tentative topics include classification, resampling methods, linear models with regularization (e.g. LASSO), Bayesian variable selection, additive models, classification and regression trees, random forests, support vector machines, basic neural networks, and basics of unsupervised learning, such as k-mean clustering, hierarchical clustering, associative rules, factor analysis, nonnegative matrix factorization.
We will discuss the application of machine learning methods to genomics, which will include (but not limited to) variant annotation, genetic association analysis, fine mapping, risk prediction, variant calling and filtering from next generation sequencing data.
We will illustrate examples with Python (scikit learn). Classical machine learning packages are also wrapped in R (mlr) and students are encouraged to learn R as well.
- Introduction (ppt) (08/23 08/25)
- Linear models (ppt) (Notes: 08/25 08/30 09/08 09/13 09/15 09/20)
- Introduction to statistical genetics (ppt)
- From GWAS to biology (notes)
- From GWAS to clinical insights (notes)
- Linear discriminant analysis (ppt) (notes: 10/11 10/13 10/25+27)
- Support vector machine (ppt) (notes: 11/01 11/03 11/07 11/10)
- Model assessment (ppt) (notes: 11/10+15 11/17 11/29)
- Classification and regression tree (ppt) (notes: 11/29 12/03+05)
- Random forest (ppt) (notes: 12/06)
- Boosting (ppt) (notes: 12/08)