AM207 - Stochastic Methods for Data Analysis, Inference and Optimization

Logo

This is a introductory graduate course on probabilistic modeling and inference for machine learning.

Course Project Description

The aim of the project is to acclimate you to the process of conducting research in statistical modeling or machine learning: 1) tackle an intimidatingly general problem 2) narrow the scope to a specific set of parameters to study 3) replicate existing literature studying these parameters 4) innovate on top of existing work incrementally, and 5) consider the wider socio-technological impact of your work.

Students who make meaningful novel or insightful contributions in their course projects have the opportunity to develop their project into a workshop or conference publication during the Spring 2022 semester. Examples of publications from AM207 Fall 2021:

Due:

    1. (October 2nd) Project Checkpoint #1: Team Formation
      • Submit the names of your team members
    2. (October 23rd) Project Checkpoint #2: Paper Selection
      • Meet with the instructor to discuss paper choice prior to this date
      • Submit the name of your paper
    3. (November 6th) Project Checkpoint #3: Paper Overview
      • Meet with your assigned TF prior to this date to discuss your paper
      • Submit a summary of the high-level ideas of your paper 
    4. (November 20th) Project Checkpoint #4: Pedagogical Examples
      • Meet with the instructor prior to this date to discuss plans for implementation and pedagogical examples
      • Show implementation/experimentation in progress
    5. (December 17th) Final Deliverable Due

Grading: Your grade is based on:

  1.  How well your deliverable satisfies the project specifications (described below)
  2. Your progress leading up to the final deliverable. There will be a series of check points with your project TF and instructor through out the semester. Your final grade will account for how much progress you've made by each check point.

Instructions:

Your exposition should focus on summarization and highlighting (aiming for an audience of peers who have taken AM207). There is no point rewording the paper itself. Reorganize and explain the ideas in a way that makes sense to you, that features the most salient/important aspects of the paper, that demonstrates understanding and synthesis.

You are welcome to study any code that is provided with the paper, you are however not allowed to copy code. Your implementation must be your own. If a public repo is available for your paper, you are encouraged to first try reproducing some results using the authors code -- this will give you an idea of how their algorithm/model works.

Examples of Project from Fall 2019:

Github repo of Fall 2019 AM207 projects

List of Pre-Approved Papers: 

Out of Distribution Detection, Dataset Shifts

  1. Know Your Limits: Uncertainty Estimation with ReLU Classifiers Fails at Reliable OOD Detection
  2. Identifying Untrustworthy Predictions in Neural Networks by Geometric Gradient Analysis
  3. The Dimpled Manifold Model of Adversarial Examples in Machine Learning
  4. Adversarial Examples are not Bugs, they are FeaturesA Discussion of Adversarial Examples Are Not Bugs, They Are Features 
  5. Wilds: A Benchmark of in-the-Wild Distribution Shifts
  6. CSI: Novelty Detection via Contrastive Learning on Distributionally Shifted Instances
  7. Provably Robust Detection of Out-of-distribution Data (almost) for free 

Interpretations, Explanations, Fairness of Models

  1. Local Explanations via Necessity and Sufficiency: Unifying Theory and Practice 
  2. BayLIME: Bayesian Local Interpretable Model-Agnostic Explanations BayLIME: Bayesian Local Interpretable Model-Agnostic Explanations
  3. Explaining Bayesian Neural Networks
  4. Measuring Data Leakage in Machine-Learning Models with Fisher Information
  5. Have We Learned to Explain?: How Interpretability Methods Can Learn to Encode Predictions in their Interpretations
  6. Saliency is a Possible Red Herring When Diagnosing Poor Generalization
  7. Selective Classification Can Magnify Disparities Across Groups
  8. Counterfactual risk assessments, evaluation, and fairness
  9. Does knowledge distillation really work?
  10. Feature Attributions and Counterfactual Explanations Can Be Manipulated
  11. POTs: Protective Optimization Technologies 
  12. Reliable Post hoc Explanations: Modeling Uncertainty in Explainability
  13. Removing Spurious Features can Hurt Accuracy and Affect Groups Disproportionately

Impacts and Applications of ML

  1. Learning advanced mathematical computations from examples
  2. The effect of differential victim crime reporting on predictive policing systems; Beyond Bias: Re-Imagining the Terms of ‘Ethical AI’ in Criminal Law 
  3. A Bayesian Model of Cash Bail Decisions
  4. Phenotyping with Prior Knowledge using Patient Similarity
  5. Hidden Risks of Machine Learning Applied to Healthcare: Unintended Feedback Loops Between Models and Future Data Causing Model Degradation
  6. Risk score learning for COVID-19 contact tracing apps
  7. CheXbreak: Misclassification Identification for Deep Learning Models Interpreting Chest X-rays

Deep Bayesian and Ensemble Models

  1. Variational Refinement for Importance Sampling Using the Forward Kullback-Leibler Divergence
  2. Diagnostics for Conditional Density Models and Bayesian Inference Algorithms
  3. Post-hoc loss-calibration for Bayesian neural networks 
  4. Scaling Hamiltonian Monte Carlo Inference for Bayesian Neural Networks with Symmetric Splitting 
  5. Learnable Uncertainty under Laplace Approximations 
  6. Rethinking Function-Space Variational Inference in Bayesian Neural Networks
  7. Fast Adaptation with Linearized Neural Networks 
  8. Beyond Marginal Uncertainty: How Accurately can Bayesian Regression Models Estimate Posterior Predictive Correlations?
  9. Bayesian Deep Learning via Subnetwork Inference
  10. What Are Bayesian Neural Network Posteriors Really Like?
  11. Loss Surface Simplexes for Mode Connecting Volumes and Fast Ensembling
  12. No MCMC for Me: Amortized Sampling for Fast and Stable Training of Energy-Based Models
  13. Repulsive Deep Ensembles are Bayesian Repulsive Deep Ensembles are Bayesian
  14. Learning Neural Network Subspaces
  15. Deep Ensemble Uncertainty Fails as Network Width Increases: Why, and How to Fix It