CS620: New Trends in IT (Formal Methods in Machine Learning)
|
Instructor | Supratik Chakraborty | email: supratik [AT] cse ... |
|
Machine learning (ML) is increasingly being used in applications
where formal guarantees are desirable. For example, self-driven
vehicles use ML components, wrong decisions by which can lead to fatal
accidents. The ACAS Xu system for collision avoidance of autonomous
drones is another such example, where a neural network is required to
satisfy properties rigorously specified in logic. Formal robustness
guarantees of image classification networks require the class label to
not change unless the image is perturbed beyond an acceptable
threshold. For various historical reasons, mainstream ML research has
focused on development of an extremely rich set of tools, techiniques
and models that can be used to solve an array of real-world problems,
without formal guarantees of how they behave in all possible corner
cases. In this course, we will be primarily interested in the
behaviour of ML components in these hard-to-fish-out corner case
scenarios. We will study how techniques from formal methods (FM) can
be used to provide rigorous guarantees of the behaviour of ML
components like deep neural networks (DNN) and recurrent neural
networks (RNN). Formal methods have been successfully used to provide
rigorous guarantees for hardware and software systems in the past,
related work leading to at least three Turing Awards. The application
of FM to prove properties of modern ML components, however, presents
new challenges due to the scale and size of these ML components, and
also because of the specifics of their operation. Over the last
decade, this has spurred a lot of interest from both the ML and FM
research communities, and formal methods for machine learning is now a
routine track in most ML/FM conferences. This course will primarily
focus on four threads of research in this area: (i) use of abstraction
refinement for analyzing DNNs, (ii) use of specialized constraint
solving techniques for reasoning about DNNs, (iii) automata theoretic
analysis of RNNs, and (iv) formal methods for explainability and
robustness of ML components. The course should be useful for students
of both formal methods and machine learning, and lies at the
intersection of these areas. Since the area is relatively new, the
course material will be primarily based on research papers.
|
UG students | CS228, CS213, CS337 |
PG students | Either CS433 or CS719 or CS771. CS725 is recommended but not essential |
Pre-recorded lectures will be posted by Tues 5.30pm (exceptions to be announced) |
Interactive doubt-clearing sessions on MSTeams CS620 channel, Mon and Thurs, 5.30-7.00 pm |
Midsem | 30% |
Endsem | 30% |
Paper critique presentation | 10% |
Project | 20% |
Quizes (attention-recall type) | 10% |
|
|
Week 2 |
|
Week 3 |
|
Week 4 |
|
Week 5 |
|
Week 6 |
|
Week 7 |
|
Weeks 8-9 |
|
Weeks 10-11 |
|
Week 12 |
|
Week 13 |
|
Week 14 |
|
Exam/Quiz | Posted on | Answers due by |
---|---|---|
Quiz 1 | Jan 29, 11:59am | Jan 31, 11:59am (upload PDF file on Moodle) |
Quiz 2 | Feb 21, 3:00pm | Feb 22, 5:00pm (upload PDF file on Moodle) |
Midsem Exam | Feb 24, 6:10pm | Feb 24, 8:10pm (upload PDF file on Moodle) |
Quiz 3 | Apr 18, 3:30pm | Apr 18, 7:30pm (upload PDF file on Moodle) |
Endsem Exam | May 1, 10:00am | May 1, 4:00pm (upload PDF file on Moodle) |
Endsem Re-exam | May 13, 3:30pm | May 13, 9:30pm (upload PDF file on Moodle) |
Week | Topic | Video recordings | Slides |
---|---|---|---|
Week 1 (Jan 7) | Course Logistics | Interaction session: Part 1 (Lectures and timings) Part 2 (What the course is/isn't about) Part 3 (Expectations from students) |
PDF slides |
Week 2 (Jan 11 - 18) | What, Why and How of FM in ML | Pre-recorded lectures: Part 1 (Safety in AI/ML -- ML perspective) Part 2 (Safety in AI/ML -- FM perspective) Part 3 (Safety in AI/ML -- Delving deeper from FM perspective) Salient points from interaction session on Jan 11, 2021 included in above videos |
PDF slides |
Week 3 (Jan 19 - 25) | Specifying Properties of Neural Networks | Pre-recorded lectures: Part 1 (Hoare triples for specifying properties of neural networks) Part 2 (Pitfalls of specifying properties of neural neworks) Part 3 (Non-perceptual and perceptual property specifications -- issues and challenges) Part 4 (Specs for perceptual DNNs, writing multiple sub-specs and system-level specs) |
PDF slides |
Week 4 (Jan 26 - Feb 1) | Catching-up week (and Quiz 1) | No pre-recorded lectures | No PDF slides |
Week 5 (Feb 1 - 8) | Modeling Neural Networks | Pre-recorded lectures: Part 1 (Modeling nodes and edges as first-order constraints) Part 2 (Compositional model building of entire network) Part 3 (Relational models and approximations) Part 4 (Approximate systems of constraints) |
PDF slides |
Week 6 (Feb 9 - 16) | Specialized decision procedures for reasoning about deep neural networks | Pre-recorded lectures: Talk on Reluplex by Guy Katz, the lead author of the Reluplex paper. |
No PDF slides |
Week 7 (Feb 16 - Feb 22) | Catching-up week, revision for mid-sem and Quiz 2 | No pre-recorded lectures | No PDF slides |
Week 8 (Mar 4 - Mar 11) | Discussion on midsem Abstract interpretation for reasoning about neural networks |
Interaction session: Part 1 (Discussion on midsem question 1) Part 2 (Discussion on midsem question 2) Part 3 (Glimpse of next topic, discussion on projects) Pre-recorded lectures Part 1 (Primer on Abstract Interpretation: Abstract domains) Part 2 (Primer on Abstract Interpretation: Abstract state transitions) |
PDF slides (primer on abstract interpretation) |
Week 9 (Mar 12 - Mar 18) | Continuing discussion on DeepPoly: Polyhedral abstract interpretation for neural networks | Interaction session: Edited recording coming soon Pre-recorded lectures Talk on DeepPoly by Martin Vechev, co-author of the DeepPoly paper |
No PDF slides |
Weeks 10-11 (Mar 19 - Apr 1) | Training networks that are verifiably "correct" | Pre-recorded lectures |
|
Week 12 (Apr 2 - Apr 9) | Recurrent neural networks and automata | Pre-recorded lectures |
|
Week 13 (Apr 10 - Apr 15) | Safety in Reinforcement Learning | Pre-recorded lectures |
|
Week 14 (Apr 16 - Apr 23) | Endsem revision and Quiz 3 | Pre-recorded lectures |