Skip to main content

This lecture and tutorial focuses on measuring human functional brain networks, as well as how to account for inherent variability within those networks. 

Difficulty level: Intermediate
Duration: 50:44
Speaker: : Caterina Gratton

This lesson provides an overview of Jupyter notebooks, Jupyter lab, and Binder, as well as their applications within the field of neuroimaging, particularly when it comes to the writing phase of your research. 

Difficulty level: Intermediate
Duration: 50:28
Speaker: : Elizabeth DuPre

This lesson provides a brief introduction to the Computational Modeling of Neuronal Plasticity.

Difficulty level: Intermediate
Duration: 0:40

In this lesson, you will be introducted to a type of neuronal model known as the leaky integrate-and-fire (LIF) model.

Difficulty level: Intermediate
Duration: 1:23

This lesson goes over various potential inputs to neuronal synapses, loci of neural communication.

Difficulty level: Intermediate
Duration: 1:20

This lesson describes the how and why behind implementing integration time steps as part of a neuronal model.

Difficulty level: Intermediate
Duration: 1:08

In this lesson, you will learn about neural spike trains which can be characterized as having a Poisson distribution.

Difficulty level: Intermediate
Duration: 1:18

This lesson covers spike-rate adaptation, the process by which a neuron's firing pattern decays to a low, steady-state frequency during the sustained encoding of a stimulus.

Difficulty level: Intermediate
Duration: 1:26

This lesson provides a brief explanation of how to implement a neuron's refractory period in a computational model.

Difficulty level: Intermediate
Duration: 0:42

In this lesson, you will learn a computational description of the process which tunes neuronal connectivity strength, spike-timing-dependent plasticity (STDP).

Difficulty level: Intermediate
Duration: 2:40

This lesson reviews theoretical and mathematical descriptions of correlated spike trains.

Difficulty level: Intermediate
Duration: 2:54

This lesson investigates the effect of correlated spike trains on spike-timing dependent plasticity (STDP).

Difficulty level: Intermediate
Duration: 1:43

This lesson goes over synaptic normalisation, the homeostatic process by which groups of weighted inputs scale up or down their biases.

Difficulty level: Intermediate
Duration: 2:58

In this lesson, you will learn about the intrinsic plasticity of single neurons.

Difficulty level: Intermediate
Duration: 2:08

This lesson covers short-term facilitation, a process whereby a neuron's synaptic transmission is enhanced for a short (sub-second) period.

Difficulty level: Intermediate
Duration: 1:58

This lesson describes short-term depression, a reduction of synaptic information transfer between neurons.

Difficulty level: Intermediate
Duration: 1:40

This lesson briefly wraps up the course on Computational Modeling of Neuronal Plasticity.

Difficulty level: Intermediate
Duration: 0:37

This is the Introductory Module to the Deep Learning Course at CDS, a course that covered the latest techniques in deep learning and representation learning, focusing on supervised and unsupervised deep learning, embedding methods, metric learning, convolutional and recurrent nets, with applications to computer vision, natural language understanding, and speech recognition.

Difficulty level: Intermediate
Duration: 50:17

This module covers the concepts of gradient descent and the backpropagation algorithm and is a part of the Deep Learning Course at NYU's Center for Data Science.

Difficulty level: Intermediate
Duration: 1:51:03
Speaker: : Yann LeCun

This lecture covers concepts associated with neural nets, including rotation and squashing, and is a part of the Deep Learning Course at New York University's Center for Data Science (CDS).

Difficulty level: Intermediate
Duration: 1:01:53
Speaker: : Alfredo Canziani