This lesson provides an introduction to the lifecycle of EEG/ERP data, describing the various phases through which these data pass, from collection to publication.
In this lesson you will learn about experimental design for EEG acquisition, as well as the first phases of the EEG/ERP data lifecycle.
This lesson provides an overview of the current regulatory measures in place regarding experimental data security and privacy.
In this lesson, you will learn the appropriate methods for collection of both data and associated metadata during EEG experiments.
This lesson goes over methods for managing EEG/ERP data after it has been collected, from annotation to publication.
In this final lesson of the course, you will learn broadly about EEG signal processing, as well as specific applications which make this kind of brain signal valuable to researchers and clinicians.
This lesson describes the principles underlying functional magnetic resonance imaging (fMRI), diffusion-weighted imaging (DWI), tractography, and parcellation. These tools and concepts are explained in a broader context of neural connectivity and mental health.
This tutorial introduces pipelines and methods to compute brain connectomes from fMRI data. With corresponding code and repositories, participants can follow along and learn how to programmatically preprocess, curate, and analyze functional and structural brain data to produce connectivity matrices.
This lesson introduces the practical exercises which accompany the previous lessons on animal and human connectomes in the brain and nervous system.
This lecture and tutorial focuses on measuring human functional brain networks, as well as how to account for inherent variability within those networks.
This lecture presents an overview of functional brain parcellations, as well as a set of tutorials on bootstrap agregation of stable clusters (BASC) for fMRI brain parcellation.
This lecture covers a wide range of aspects regarding neuroinformatics and data governance, describing both their historical developments and current trajectories. Particular tools, platforms, and standards to make your research more FAIR are also discussed.
This lecture introduces you to the basics of the Amazon Web Services public cloud. It covers the fundamentals of cloud computing and goes through both the motivations and processes involved in moving your research computing to the cloud.
This lecture discusses how FAIR practices affect personalized data models, including workflows, challenges, and how to improve these practices.
In this talk, you will learn how brainlife.io works, and how it can be applied to neuroscience data.
As a part of NeuroHackademy 2020, this lecture delves into cloud computing, focusing on Amazon Web Services.
This talk presents an overview of CBRAIN, a web-based platform that allows neuroscientists to perform computationally intensive data analyses by connecting them to high-performance computing facilities across Canada and around the world.
This lesson continues from part one of the lecture Ontologies, Databases, and Standards, diving deeper into a description of ontologies and knowledg graphs.
This lecture covers FAIR atlases, including their background and construction, as well as how they can be created in line with the FAIR principles.
This lecture focuses on ontologies for clinical neurosciences.