Instructor : Cem Subakan

This class is about learning to build machine learning algorithms for signals. Different from a standard machine learning class, we will have a little bit more of an EE flavor to things. That is, we will often work with sequential data such as speech and audio, and other signals. We will give the necessary background to be able to propose and carry out a research or applied project in the domain of machine learning for signal processing. In the end, our goal is to teach how to fish (for MLSP projects)!

This class is influenced by classes with the same title in UIUC, CMU, and Indiana University.

Week 1 : Linear Algebra Refresher slides, lab

Matrix multiplication

Index, Matrix, Tensor notations

Eigenvalues, Eigenvectors

Building the reflexes to avoid for loops

Signal representations

Tensors, Funky Tensor Mathematics

Linear Algebraic Matrix Decompositions

Week 2 : Probability Refresher slides, lab

Probability Calculus, Bayes Rule

Continuous and Discrete Random Variables

Multidimensional Random Variables

Probabilistic Graphical Model Conventions

Week 3: Signal Processing Refresher slides, lab

Continuous and Discrete Signals

Sampling, Analog to Digital Conversion

Fourier Transform, Discrete-Cosine Transform, Short Time Fourier Transform

Filtering

Mechnanics of Convolution in Time Domain, Convolution as a Matrix Multiply

Week 4: Machine Learning 1: Decompositions slides, lab

Linear Regression

Linear Regression connections with Fourier Transform

Dimensionality Reduction, PCA and its variants, ICA, NMF

Week 5: Machine Learning 2: Non-linear Dimensionality Reduction slides, lab

Kernel PCA

Multidimensional Scaling

Manifold Learning Methods

ISOMAP

Locally Linear Embeddings

Laplacian Eigenmaps

TSNE

Week 6: Machine Learning 3: Classification slides, lab

Generative Classification

Discriminative Classification

Perceptron Algorithm

Logistic Regression

Kernel Logistic Regression

Neural Network Classifier

Week 7: Deep Learning Primer slides, lab

Feedforward Networks

Skip Connections

Convolutional Layers

Recurrent Layers

Attention Layers

Gradient Descent and variants

Week 8 (Invited Lecture by Sara Karami) slides

Week 9: Machine Learning 4: Clustering slides, lab

Kmeans clustering

Mixture Models

Expectation Maximization, Iterative Conditional Modes

Spectral Clustering

Hierarchical Clustering

Week 10: Time Series Models slides, lab

Dynamic Time Warping

Hidden Markov Models, Forward-Backward Algorithm

EM for HMMs

Viterbi Decoding

HMM Variants (Mixture of HMMs, Factorial HMMs,…)

Week 11: Graph Signal Processing / Graph ML slides, lab

Signals as Graphs

Graph Fourier Transform

Graph Methods for Signal Processing

Graph Convolutions

Graph Neural Networks

Week 12: Speech / Audio slides

Automatic Speech Recognition (ASR)

Text-to-Speech

Speech Separation / Enhancement

Interpretability in the Audio Domain

Text-Audio Multi-Modal Representations

Week 13: Project Presentations

There will be 3 homeworks (45%), labs (10%) and a project (45%) that will be carried out by teams of 2-3 students. It is preferable that the students propose the project, but we will propose several projects ideas also.