Skip to main content

In this tutorial on simulating whole-brain activity using Python, participants can follow along using corresponding code and repositories, learning the basics of neural oscillatory dynamics, evoked responses and EEG signals, ultimately leading to the design of a network model of whole-brain anatomical connectivity. 

Difficulty level: Intermediate
Duration: 1:16:10
Speaker: : John Griffiths

Tutorial on how to simulate brain tumor brains with TVB (reproducing publication: Marinazzo et al. 2020 Neuroimage). This tutorial comprises a didactic video, jupyter notebooks, and full data set for the construction of virtual brains from patients and health controls. Authors: Hannelore Aerts, Michael Schirner, Ben Jeurissen, DIrk Van Roost, Eric Achten, Petra Ritter, Daniele Marinazzo

Difficulty level: Intermediate
Duration: 10:01
Speaker: :

The tutorial comprises a didactic video and jupyter notebooks (reproducing publication: Falcon et al. 2016 eNeuro). Contributors: Daniele Marinazzo, Petra Ritter, Paul Triebkorn, Ana Solodkin

Difficulty level: Intermediate
Duration: 7:43
Speaker: :

Learn how to create a standard extracellular electrophysiology dataset in NWB using Python

Difficulty level: Intermediate
Duration: 23:10
Speaker: : Ryan Ly

Learn how to create a standard calcium imaging dataset in NWB using Python

Difficulty level: Intermediate
Duration: 31:04
Speaker: : Ryan Ly

Learn how to create a standard intracellular electrophysiology dataset in NWB

Difficulty level: Intermediate
Duration: 20:23
Speaker: : Pamela Baker

Learn how to create a standard intracellular electrophysiology dataset in NWB

Difficulty level: Intermediate
Duration: 20:22
Speaker: : Pamela Baker
Course:

This book was written with the goal of introducing researchers and students in a variety of research fields to the intersection of data science and neuroimaging. This book reflects our own experience of doing research at the intersection of data science and neuroimaging and it is based on our experience working with students and collaborators who come from a variety of backgrounds and have a variety of reasons for wanting to use data science approaches in their work. The tools and ideas that we chose to write about are all tools and ideas that we have used in some way in our own research. Many of them are tools that we use on a daily basis in our work. This was important to us for a few reasons: the first is that we want to teach people things that we ourselves find useful. Second, it allowed us to write the book with a focus on solving specific analysis tasks. For example, in many of the chapters you will see that we walk you through ideas while implementing them in code, and with data. We believe that this is a good way to learn about data analysis, because it provides a connecting thread from scientific questions through the data and its representation to implementing specific answers to these questions. Finally, we find these ideas compelling and fruitful. That’s why we were drawn to them in the first place. We hope that our enthusiasm about the ideas and tools described in this book will be infectious enough to convince the readers of their value.

 

Difficulty level: Intermediate
Duration:
Speaker: :
Course:

This Jupyter Book is a series of interactive tutorials about quantitative T1 mapping, powered by qMRLab. Most figures are generated with Plot.ly – you can play with them by hovering your mouse over the data, zooming in (click and drag) and out (double click), moving the sliders, and changing the drop-down options. To view the code that was used to generate the figures in this blog post, hover your cursor in the top left corner of the frame that contains the tutorial and click the checkbox “All cells” in the popup that appears.

Jupyter Lab notebooks of these tutorials are also available through MyBinder, and inline code modification inside the Jupyter Book is provided by Thebelab. For both options, you can modify the code, change the figures, and regenerate the html that was used to create the tutorial below. This Jupyter Book also uses a Script of Scripts (SoS) kernel, allowing us to process the data using qMRLab in MATLAB/Octave and plot the figures with Plot.ly using Python, all within the same Jupyter Notebook.

Difficulty level: Intermediate
Duration:
Speaker: :

This lecture introduces neuroscience concepts and methods such as fMRI, visual respones in BOLD data, and the eccentricity of visual receptive fields. 

Difficulty level: Intermediate
Duration: 7:15
Speaker: : Mike X. Cohen

This tutorial walks users through the creation and visualization of activation flat maps from fMRI datasets. 

Difficulty level: Intermediate
Duration: 12:15
Speaker: : Mike X. Cohen

This tutorial demonstrates to users the conventional preprocessing steps when working with BOLD signal datasets from fMRI. 

Difficulty level: Intermediate
Duration: 12:05
Speaker: : Mike X. Cohen

In this tutorial, users will learn how to create a trial-averaged BOLD response and store it in a matrix in MATLAB. 

Difficulty level: Intermediate
Duration: 20:12
Speaker: : Mike X. Cohen

This tutorial teaches users how to create animations of BOLD responses over time, to allow researchers and clinicians to visualize time-course activity patterns.

Difficulty level: Intermediate
Duration: 12:52
Speaker: : Mike X. Cohen

This tutorial demonstrates how to use MATLAB to create event-related BOLD time courses from fMRI datasets. 

Difficulty level: Intermediate
Duration: 13:39
Speaker: : Mike X. Cohen

In this tutorial, users learn how to compute and visualize a t-test on experimental condition differences.

Difficulty level: Intermediate
Duration: 17:54
Speaker: : Mike X. Cohen

You will learn about working with calcium imaging data, including image processing to remove background "blur," identifying cells based on thresholded spatial contiguity, time series filtering, and principal components analysis (PCA). The MATLAB code shows data animations, capabilities of the image processing toolbox, and PCA.

Difficulty level: Intermediate
Duration: 5:02
Speaker: : Mike X. Cohen

You will learn about working with calcium imaging data, including image processing to remove background "blur," identifying cells based on thresholded spatial contiguity, time series filtering, and principal components analysis (PCA). The MATLAB code shows data animations, capabilities of the image processing toolbox, and PCA.

Difficulty level: Intermediate
Duration: 15:01
Speaker: : Mike X. Cohen

You will learn about working with calcium imaging data, including image processing to remove background "blur," identifying cells based on thresholded spatial contiguity, time series filtering, and principal components analysis (PCA). The MATLAB code shows data animations, capabilities of the image processing toolbox, and PCA.

Difficulty level: Intermediate
Duration: 5:15
Speaker: : Mike X. Cohen

You will learn about working with calcium imaging data, including image processing to remove background "blur," identifying cells based on thresholded spatial contiguity, time series filtering, and principal components analysis (PCA). The MATLAB code shows data animations, capabilities of the image processing toolbox, and PCA.

Difficulty level: Intermediate
Duration: 17:08
Speaker: : Mike X. Cohen