This is a tutorial covering Generalized Linear Models (GLMs), which are a fundamental framework for supervised learning. In this tutorial, the objective is to model a retinal ganglion cell spike train by fitting a temporal receptive field: first with a Linear-Gaussian GLM (also known as ordinary least-squares regression model) and then with a Poisson GLM (aka "Linear-Nonlinear-Poisson" model). The data you will be using was published by Uzzell & Chichilnisky 2004.
This tutorial covers multivariate data can be represented in different orthonormal bases.
This tutorial covers how to perform principal component analysis (PCA) by projecting the data onto the eigenvectors of its covariance matrix.
To quickly refresh your knowledge of eigenvalues and eigenvectors, you can watch this short video (4 minutes) for a geometrical explanation. For a deeper understanding, this in-depth video (17 minutes) provides an excellent basis and is beautifully illustrated.
This tutorial covers how to apply principal component analysis (PCA) for dimensionality reduction, using a classic dataset that is often used to benchmark machine learning algorithms: MNIST. We'll also learn how to use PCA for reconstruction and denoising.
You can learn more about MNIST dataset here.
This tutorial covers how dimensionality reduction can be useful for visualizing and inferring structure in your data. To do this, we will compare principal component analysis (PCA) with t-SNE, a nonlinear dimensionality reduction method.
This lecture introduces the concept of Bayesian statistics and explains why Bayesian statistics are relevant to studying the brain.
This tutorial provides an introduction to Bayesian statistics and covers developing a Bayesian model for localizing sounds based on audio and visual cues. This model will combine prior information about where sounds generally originate with sensory information about the likelihood that a specific sound came from a particular location. The resulting posterior distribution not only allows us to make optimal decision about the sound's origin, but also lets us quantify how uncertain that decision is. Bayesian techniques are therefore useful normative models: the behavior of human or animal subjects can be compared against these models to determine how efficiently they make use of information.
In this tutorial, we will use the concepts introduced in Tutorial 1 as building blocks to explore more complicated sensory integration and ventriloquism!
This tutorial covers computing all the necessary steps to perform model inversion (estimate the model parameters such as 𝑝𝑐𝑜𝑚𝑚𝑜𝑛 that generated data similar to that of a participant). We will describe all the steps of the generative model first, and in the last exercise we will use all these steps to estimate the parameter 𝑝𝑐𝑜𝑚𝑚𝑜𝑛 of a single participant using simulated data. The generative model will be the same Bayesian model we have been using throughout tutorial 2: a mixture of Gaussian prior (common + independent priors) and a Gaussian likelihood.
This tutorial focuses on Bayesian Decision Theory, which combines the posterior with cost functions that allow us to quantify the potential impact of making a decision or choosing an action based on that posterior. Cost functions are therefore critical for turning probabilities into actions!
This lecture focuses on advanced uses of Bayesian statistics for understanding the brain.
This lecture provides an introduction to linear systems.
This tutorial covers the behavior of dynamical systems, systems that evolve in time, where the rules by which they evolve in time are described precisely by a differential equation.
Differential equations are equations that express the rate of change of the state variable 𝑥. One typically describes this rate of change using the derivative of 𝑥 with respect to time (𝑑𝑥/𝑑𝑡) on the left hand side of the differential equation: 𝑑𝑥𝑑𝑡=𝑓(𝑥). A common notational short-hand is to write 𝑥˙ for 𝑑𝑥𝑑𝑡. The dot means "the derivative with respect to time".
This tutorial provides an introduction to the Markov process in a simple example where the state transitions are probabilistic. The aims of this tutorial is to help you understand Markov processes and history dependence, as well as to explore the behavior of a two-state telegraph process and understand how its equilibrium distribution is dependent on its parameters.
This tutorial builds on how deterministic and stochastic processes can both be a part of a dynamical system by simulating random walks, investigating the mean and variance of a Ornstein-Uhlenbeck (OU) process, and quantifying the OU process's behavior at equilibrium.
The goal of this tutorial is to use the modeling tools and intuitions developed in the previous few tutorials and use them to fit data. The concept is to flip the previous tutorial -- instead of generating synthetic data points from a known underlying process, what if we are given data points measured in time and have to learn the underlying process?
This tutorial is in two sections:
This lecture provides a summary of concepts associated with linear dynamical systems, covered in Linear Systems I (Intro Lecture) and Tutorials 1 - 4, and also introduces motor neuroscience/neuroengineering, brain-machine interfaces, and applications of dynamical systems.
This lesson provides an introduction to the Decision Making course, specifically focusing on hidden states in neural systems.
This tutorial introduces the Sequential Probability Ratio Test between two hypotheses 𝐻𝐿 and 𝐻𝑅 by running simulations of a Drift Diffusion Model (DDM). As independent and identically distributed (i.i.d) samples from the true data-generating distribution coming in, we accumulate our evidence linearly until a certain criterion is met before deciding which hypothesis to accept. Two types of stopping criterion/stopping rule will be implemented: after seeing a fixed amount of data, and after the likelihood ratio passes a pre-defined threshold. Due to the noisy nature of observations, there will be a drifting term governed by expected mean output and a diffusion term governed by observation noise.
This tutorial covers how to simulate a Hidden Markov Model (HMM) and observe how changing the transition probability and observation noise impacts what the samples look like. Then we'll look at how uncertainty increases as we make future predictions without evidence (from observations) and how to gain information from the observations.