Skip to main content

This presentation discusses the impact of data sharing in stroke.

Difficulty level: Intermediate
Duration: 16:33
Speaker: : Valeria Caso

This talks presents an overview of the potential for data federation in stroke research.

Difficulty level: Intermediate
Duration: 21:37

This lesson provides an introduction to the lifecycle of EEG/ERP data, describing the various phases through which these data pass, from collection to publication.

Difficulty level: Beginner
Duration: 35:30

In this lesson you will learn about experimental design for EEG acquisition, as well as the first phases of the EEG/ERP data lifecycle. 

Difficulty level: Beginner
Duration: 30:04

This lesson provides an overview of the current regulatory measures in place regarding experimental data security and privacy. 

Difficulty level: Beginner
Duration: 31:00

In this lesson, you will learn the appropriate methods for collection of both data and associated metadata during EEG experiments.

Difficulty level: Beginner
Duration: 29:14

This lesson goes over methods for managing EEG/ERP data after it has been collected, from annotation to publication. 

Difficulty level: Beginner
Duration: 39:25

In this final lesson of the course, you will learn broadly about EEG signal processing, as well as specific applications which make this kind of brain signal valuable to researchers and clinicians. 

Difficulty level: Beginner
Duration: 34:51

This lesson is a general overview of overarching concepts in neuroinformatics research, with a particular focus on clinical approaches to defining, measuring, studying, diagnosing, and treating various brain disorders. Also described are the complex, multi-level nature of brain disorders and the data associated with them, from genes and individual cells up to cortical microcircuits and whole-brain network dynamics. Given the heterogeneity of brain disorders and their underlying mechanisms, this lesson lays out a case for multiscale neuroscience data integration.

Difficulty level: Intermediate
Duration: 1:09:33
Speaker: : Sean Hill

This lesson breaks down the principles of Bayesian inference and how it relates to cognitive processes and functions like learning and perception. It is then explained how cognitive models can be built using Bayesian statistics in order to investigate how our brains interface with their environment. 

This lesson corresponds to slides 1-64 in the PDF below. 

Difficulty level: Intermediate
Duration: 1:28:14

This lecture and tutorial focuses on measuring human functional brain networks, as well as how to account for inherent variability within those networks. 

Difficulty level: Intermediate
Duration: 50:44
Speaker: : Caterina Gratton

This tutorial demonstrates how to work with neuronal data using MATLAB, including actional potentials and spike counts, orientation tuing curves in visual cortex, and spatial maps of firing rates.

Difficulty level: Intermediate
Duration: 5:17
Speaker: : Mike X. Cohen

This lesson instructs users on how to import electrophysiological neural data into MATLAB, as well as how to convert spikes to a data matrix.

Difficulty level: Intermediate
Duration: 11:37
Speaker: : Mike X. Cohen

In this lesson, users will learn about human brain signals as measured by electroencephalography (EEG), as well as associated neural signatures such as steady state visually evoked potentials (SSVEPs) and alpha oscillations. 

Difficulty level: Intermediate
Duration: 8:51
Speaker: : Mike X. Cohen

This lesson provides a brief overview of the Python programming language, with an emphasis on tools relevant to data scientists.

Difficulty level: Beginner
Duration: 1:16:36
Speaker: : Tal Yarkoni
Course:

In this lesson, users can follow along as a spaghetti script written in MATLAB is turned into understandable and reusable code living happily in a powerful GitHub repository.

Difficulty level: Beginner
Duration: 2:08:19
Speaker: : Agah Karakuzu
Course:

This lesson gives a quick walkthrough the Tidyverse, an "opinionated" collection of R packages designed for data science, including the use of readr, dplyr, tidyr, and ggplot2.

Difficulty level: Beginner
Duration: 1:01:39
Speaker: : Thomas Mock

This lecture describes how to build research workflows, including a demonstrate using DataJoint Elements to build data pipelines.

Difficulty level: Intermediate
Duration: 47:00
Speaker: : Dimitri Yatsenko

This lecture discusses how FAIR practices affect personalized data models, including workflows, challenges, and how to improve these practices.

Difficulty level: Beginner
Duration: 13:16
Speaker: : Kelly Shen

This lecture covers how to make modeling workflows FAIR by working through a practical example, dissecting the steps within the workflow, and detailing the tools and resources used at each step.

Difficulty level: Beginner
Duration: 15:14