Course syllabus and policies: Course handout.
Teaching staff:
Please do not send the instructor or the TAs email about the class directly to their personal accounts.
Piazza: Students are encouraged to sign up Piazza to join course discussions. If your question is about the course material and doesn’t give away any hints for the homework, please post to Piazza so that the entire class can benefit from the answer.
Lecture and tutorial hours:
Time | Location | |
---|---|---|
Lecture | Thursday 10:30 AM -12:30 PM | Sydney Smith Hall, Room 1072 |
Office hours | Wednesday 9 AM - 10 AM | Zoom (https://utoronto.zoom.us/j/83036528074), Meeting ID: 830 3652 8074, Passcode: LMP1210_24 |
This course is intended for graduate students in Health Sciences to learn the basic principles of machine learning in biomedical research and to build and strengthen their computational skills of medical research. The course aims to equip students with the fundamental knowledge of machine learning (ML). During the course, the students will acquire basic computational skills and hands-on experience to deploy ML algorithms using python. The students will learn the current practices and applications of ML in medicine, and understand what ML can and cannot do for medicine. The goal of this course to establish an essential foundation for graduate students to take the first steps in computational research in medicine.
Introduction to basic principles and current practices of machine learning in biomedical research. Focus on the fundamental ML algorithms with applications in biomedical data; the application of unsupervised learning in genomic data; the application of supervised learning for medical images.
Handout | Due | |
---|---|---|
Math Diagnostics | Handout | January 18 2024 |
Assignment #1 | Handout, Data | February 1 2024 |
Assignment #2 | Handout, Data | February 15 2024 |
Assignment #3 | Handout, Data | March 7 2024 |
Final Project | Handout | February 22 2024 (Proposal) April 8 2024 (Final report) |
Lateness: Assignments and Projects will be accepted up to 3 days late, but 10% will be deducted for each day late, rounded up to the nearest day. After that, submissions will not be accepted and will receive a score of 0.
There are several generative AI tools such as ChatGPT, GPT4, Bard, Claude available for public use as general purpose chatbots. These machine learning models are incredibly powerful and are capable of providing responses to arbitrarily complex text (and sometimes image) queries. You may use generative artificial intelligence (AI) tools, including ChatGPT and GitHub Copilot, as learning aids.
Generative AI is not required to complete any aspect of this course, and we caution you to not rely entirely on these tools to complete your coursework. Instead, we recommend treating generative AI as a supplementary tool only for exploration or drafting content. Ultimately, you (and not any AI tool) are responsible for your own learning in this course, and for all the work you submit for credit. It is your responsibility to critically evaluate the content generated, and to regularly assess your own learning independent of generative AI tools. Overreliance on generative AI may give you a false sense of how much you’ve actually learned, which will lead to poor performance on the midterm test and final exam, in later courses, or in future work or studies after graduation. You will not be permitted to use generative AI on the midterm test or final exam. Remember that you (and not any AI tool) are responsible for your own learning in this course, and for the final work you submit for this assignment. This assignments and homeworks are designed to be doable without the use of generative AI, using only concepts and skills we have covered in lecture and in the course readings.
While some generative AI tools are currently available for free in Canada, please be warned that these tools have not been vetted by the University of Toronto and might not meet University guidelines or requirements for privacy, intellectual property, security, accessibility, and records retention. Generative AI may produce content which is incorrect or misleading, or inconsistent with the expectations of this course. They may even provide citations to sources that don’t exist—and submitting work with false citations is an academic offense. These tools may be subject to service interruptions, software modifications, and pricing changes during the semester.
The final work you submit for assignments in this course must be your own, and may not include any verbatim content from generative artificial intelligence (AI) tools. You may, however, use generative AI to support your work on this assignment in the following ways:
If you choose to use any generative AI tools while working on this assignment, you must acknowledge which generative AI tools you used and how you used them. It is an academic offence to not credit sources—including generative AI—in work that you submit. This acknowledgment should take the form of:
Any uses of generative AI beyond the ones listed above are not allowed, and will be considered use of an unauthorized aid, which is a form of cheating. In summary:
Suggested readings included help you understand the course material. They are not required, i.e. you are only responsible for the material covered in lecture. Most of the suggested reading listed are more advanced than the corresponding lecture, and are of interest if you want to know where our knowledge comes from or follow current frontiers of research. Please refer to the Resources section for the texts.
B = Pattern Recognition and Machine Learning (by Bishop)
DL = Deep Learning Book
MK = Information Theory, Inference, and Learning Algorithms (by MacKay)
RG = Roger Grosse’s Notes
Date | Lecture Topic | Slides | Suggested Readings | Deliverables | |
---|---|---|---|---|---|
Lecture 1 | Jan 11 | Intro to ML in medicine, nearest neighbor classifier | Slides | B: 1-2 (emphasis on 2.5.2); DL: 2-5 | |
Lecture 2 | Jan 18 | Tree-based classifier; Python tutorial for introduction to Python | Slides Colab | DL: 2, David Liu’s CSC110/111 Course Notes, Training a Classifier, MK: p. 22-36 | Math diagnostic Due |
Lecture 3 | Jan 25 | Linear methods for regression and classification; Python tutorial for evaluation methods | Slides Colab | B: 3-4; RG: Linear Regression, Linear Classifiers | |
Lecture 4 | Feb 01 | Neural networks | Slides | DL: 6-8; RG: Multilayer Perceptrons, Backpropagation (quite technical) | Assignment #1 Due |
Lecture 5 | Feb 08 | Ensemble models; Python tutorial for supervised learning practice | Slides Colab | ||
Lecture 6 | Feb 15 | Unsupervised learning for clustering: K-means, Gaussian mixture models | Slides | B: 9; RG: Mixture Models | Assignment #2 Due |
Reading Week | Feb 22 | Final project proposal Due | |||
Lecture 7 | Feb 29 | Unsupervised learning for clustering; Python tutorial for unsupervised learning practice | Slides, Colab | Jaan Altosaar’s VAE Tutorial; Introduction to VAEs (Kingma and Welling) | |
Lecture 8 | Mar 07 | Office hours for the final project | Assignment #3 Due | ||
Lecture 9 | Mar 14 | Guest Lecturer: Dr. Robert Grant Machine Learning in Oncology |
|||
Lecture 10 | Mar 21 | Advanced deep learning methods for medical image analysis | Slides | ||
Lecture 11 | Mar 28 | Term project presentation | |||
Lecture 12 | Apr 04 | Term project presentation |
Type | Name | Description |
---|---|---|
Related Textbooks | Deep Learning (Goodfellow at al., 2016) | The Deep Learning textbook is a resource intended to help students and practitioners enter the field of machine learning in general and deep learning. |
Pattern Recognition and Machine Learning (Bishop, 2006) | Classic introductory text on machine learning, less emphasis on deep learning approaches. | |
Information Theory, Inference, and Learning Algorithms (MacKay, 2003) | A good introduction textbook that combines information theory and machine learning. | |
General Framework | PyTorch | An open source deep learning platform that provides a seamless path from research prototyping to production deployment. |
Computation Platform | Colab | Colaboratory is a free Jupyter notebook environment that requires no setup and runs entirely in the cloud. |
GCE | Google Compute Engine delivers virtual machines running in Google’s innovative data centers and worldwide fiber network. | |
AWS-EC2 | Amazon Elastic Compute Cloud (EC2) forms a central part of Amazon.com’s cloud-computing platform, Amazon Web Services (AWS), by allowing users to rent virtual computers on which to run their own computer applications. |