This lesson provides a brief introduction to the Computational Modeling of Neuronal Plasticity.
In this lesson, you will be introducted to a type of neuronal model known as the leaky integrate-and-fire (LIF) model.
This lesson goes over various potential inputs to neuronal synapses, loci of neural communication.
This lesson describes the how and why behind implementing integration time steps as part of a neuronal model.
In this lesson, you will learn about neural spike trains which can be characterized as having a Poisson distribution.
This lesson covers spike-rate adaptation, the process by which a neuron's firing pattern decays to a low, steady-state frequency during the sustained encoding of a stimulus.
This lesson provides a brief explanation of how to implement a neuron's refractory period in a computational model.
In this lesson, you will learn a computational description of the process which tunes neuronal connectivity strength, spike-timing-dependent plasticity (STDP).
This lesson reviews theoretical and mathematical descriptions of correlated spike trains.
This lesson investigates the effect of correlated spike trains on spike-timing dependent plasticity (STDP).
This lesson goes over synaptic normalisation, the homeostatic process by which groups of weighted inputs scale up or down their biases.
In this lesson, you will learn about the intrinsic plasticity of single neurons.
This lesson covers short-term facilitation, a process whereby a neuron's synaptic transmission is enhanced for a short (sub-second) period.
This lesson describes short-term depression, a reduction of synaptic information transfer between neurons.
This lesson briefly wraps up the course on Computational Modeling of Neuronal Plasticity.
This lesson provides an overview of The Virtual Brain integrated workflows on EBRAINS.
This lesson walks users through the Image Processing Pipeline, an integral part of the TVB on EBRAINS integrated workflows.
This lesson gives an overview of The Virtual Brain simulator and its integration into the Human Brain Project Cloud and EBRAINS infrastructure.
In this lesson, users will get an overview of the EBRAINS integrated Fast TVB, a C implementation of TVB that is orders of magnitude faster than the original Python TVB, and capable of performing parallelizable simulations in the cloud.
In this lesson you will learn about the Bayesian Virtual Epileptic Patient (BVEP), a research use case using TVB supported on the EBRAINS infrastructure.