This lecture briefly introduces The Virtual Brain (TVB), a multi-scale, multi-modal neuroinformatics platform for full brain network simulations using biologically realistic connectivity, as well as its potential neuroscience applications (e.g., epilepsy cases).
This lecture introduces the theoretical background and foundations that led to the development of TVB, its architecture, and features of its major software components.
In this lesson, you will hear about the current challenges regarding data management, as well as policies and resources aimed to address them.
This lecture provides an overview of successful open-access projects aimed at describing complex neuroscientific models, and makes a case for expanded use of resources in support of reproducibility and validation of models against experimental data.
This lecture provides an introduction to reproducibility issues within the fields of neuroimaging and fMRI, as well as an overview of tools and resources being developed to alleviate the problem.
This lecture provides a historical perspective on reproducibility in science, as well as the current limitations of neuroimaging studies to date. This lecture also lays out a case for the use of meta-analyses, outlining available resources to conduct such analyses.
This lecture on model types introduces the advantages of modeling, provide examples of different model types, and explain what modeling is all about.
This lecture summarizes the concepts introduced in Model Types I and further explains how models can be used answer different scientific questions.
This lecture focuses on how to get from a scientific question to a model using concrete examples. We will present a 10-step practical guide on how to succeed in modeling. This lecture contains links to 2 tutorials, lecture/tutorial slides, suggested reading list, and 3 recorded Q&A sessions.
This lecture formalizes modeling as a decision process that is constrained by a precise problem statement and specific model goals. We provide real-life examples on how model building is usually less linear than presented in Modeling Practice I.
This lecture focuses on the purpose of model fitting, approaches to model fitting, model fitting for linear models, and how to assess the quality and compare model fits. We will present a 10-step practical guide on how to succeed in modeling.
This lecture summarizes the concepts introduced in Model Fitting I and adds two additional concepts: 1) MLE is a frequentist way of looking at the data and the model, with its own limitations. 2) Side-by-side comparisons of bootstrapping and cross-validation.
This lecture provides an overview of the generalized linear models (GLM) course, originally a part of the Neuromatch Academy (NMA), an interactive online summer school held in 2020. NMA provided participants with experiences spanning from hands-on modeling experience to meta-science interpretation skills across just about everything that could reasonably be included in the label "computational neuroscience".
This lecture further develops the concepts introduced in Machine Learning I. This lecture is part of the Neuromatch Academy (NMA), an interactive online computational neuroscience summer school held in 2020.
This lesson provides an overview of the process of developing the TVB-NEST co-simulation on the EBRAINS infrastructure, and its use cases.
This lecture introduces the core concepts of dimensionality reduction.
This lecture covers the application of dimensionality reduction applied to multi-dimensional neural recordings using brain-computer interfaces with simultaneous spike recordings.
This is the first of a series of tutorials on fitting models to data. In this tutorial, we start with simple linear regression, using least squares optimization.
In this tutorial, we will use a different approach to fit linear models that incorporates the random 'noise' in our data.
This tutorial discusses how to gauge how good our estimated model parameters are.