This lesson covers spike-rate adaptation, the process by which a neuron's firing pattern decays to a low, steady-state frequency during the sustained encoding of a stimulus.
This lesson provides a brief explanation of how to implement a neuron's refractory period in a computational model.
In this lesson, you will learn a computational description of the process which tunes neuronal connectivity strength, spike-timing-dependent plasticity (STDP).
This lesson reviews theoretical and mathematical descriptions of correlated spike trains.
This lesson investigates the effect of correlated spike trains on spike-timing dependent plasticity (STDP).
This lesson goes over synaptic normalisation, the homeostatic process by which groups of weighted inputs scale up or down their biases.
In this lesson, you will learn about the intrinsic plasticity of single neurons.
This lesson covers short-term facilitation, a process whereby a neuron's synaptic transmission is enhanced for a short (sub-second) period.
This lesson describes short-term depression, a reduction of synaptic information transfer between neurons.
This lesson briefly wraps up the course on Computational Modeling of Neuronal Plasticity.
This is a tutorial on how to simulate neuronal spiking in brain microcircuit models, as well as how to analyze, plot, and visualize the corresponding data.
This video will document the process of running an app on brainlife, from data staging to archiving of the final data outputs.
This quick video presents some of the various visualizers available on brainlife.io
This short video shows how a brainlife.io publication can be opened from the Data Deposition page of the journal Nature Scientific Data.
An introduction to data management, manipulation, visualization, and analysis for neuroscience. Students will learn scientific programming in Python, and use this to work with example data from areas such as cognitive-behavioral research, single-cell recording, EEG, and structural and functional MRI. Basic signal processing techniques including filtering are covered. The course includes a Jupyter Notebook and video tutorials.
This lesson breaks down the principles of Bayesian inference and how it relates to cognitive processes and functions like learning and perception. It is then explained how cognitive models can be built using Bayesian statistics in order to investigate how our brains interface with their environment.
This lesson corresponds to slides 1-64 in the PDF below.
This is a tutorial on designing a Bayesian inference model to map belief trajectories, with emphasis on gaining familiarity with Hierarchical Gaussian Filters (HGFs).
This lesson corresponds to slides 65-90 of the PDF below.
This lecture covers a lot of post-war developments in the science of the mind, focusing first on the cognitive revolution, and concluding with living machines.
This lecture provides an overview of depression (epidemiology and course of the disorder), clinical presentation, somatic co-morbidity, and treatment options.
This lesson is part 1 of 2 of a tutorial on statistical models for neural data.