This lesson describes the principles underlying functional magnetic resonance imaging (fMRI), diffusion-weighted imaging (DWI), tractography, and parcellation. These tools and concepts are explained in a broader context of neural connectivity and mental health.
This tutorial introduces pipelines and methods to compute brain connectomes from fMRI data. With corresponding code and repositories, participants can follow along and learn how to programmatically preprocess, curate, and analyze functional and structural brain data to produce connectivity matrices.
This lesson introduces the practical exercises which accompany the previous lessons on animal and human connectomes in the brain and nervous system.
This lecture and tutorial focuses on measuring human functional brain networks, as well as how to account for inherent variability within those networks.
This lecture presents an overview of functional brain parcellations, as well as a set of tutorials on bootstrap agregation of stable clusters (BASC) for fMRI brain parcellation.
This lesson provides an overview of the current status in the field of neuroscientific ontologies, presenting examples of data organization and standards, particularly from neuroimaging and electrophysiology.
This lesson continues from part one of the lecture Ontologies, Databases, and Standards, diving deeper into a description of ontologies and knowledg graphs.
This lecture covers the NIDM data format within BIDS to make your datasets more searchable, and how to optimize your dataset searches.
This lecture covers positron emission tomography (PET) imaging and the Brain Imaging Data Structure (BIDS), and how they work together within the PET-BIDS standard to make neuroscience more open and FAIR.
This lecture discusses the FAIR principles as they apply to electrophysiology data and metadata, the building blocks for community tools and standards, platforms and grassroots initiatives, and the challenges therein.
This lecture discusses how to standardize electrophysiology data organization to move towards being more FAIR.
The International Brain Initiative (IBI) is a consortium of the world’s major large-scale brain initiatives and other organizations with a vested interest in catalyzing and advancing neuroscience research through international collaboration and knowledge sharing. This workshop introduces the IBI, the efforts of the Data Standards and Sharing Working Group, and keynote lectures on the impact of data standards and sharing on large-scale brain projects, as well as a discussion on prospects and needs for neural data sharing.
This lesson gives a description of the BrainHealth Databank, a repository of many types of health-related data, whose aim is to accelerate research, improve care, and to help better understand and diagnose mental illness, as well as develop new treatments and prevention strategies.
This lesson corresponds to slides 46-78 of the PDF below.
This talk goes over Neurobagel, an open-source platform developed for improved dataset sharing and searching.
This lightning talk describes the heterogeneity of the MR field regarding types of scanners, data formats, protocols, and software/hardware versions, as well as the challenges and opportunities for unifying these datasets in a common interface, MRdataset.
This lesson describes the current state of brain-computer interface (BCI) standards, including the present obstacles hindering the forward movement of BCI standardization as well as future steps aimed at solving this problem.
This lightning talk gives an outline of the DataLad ecosystem for large-scale collaborations, and how DataLad addresses challenges that may arise in such research cooperations.
In this lightning talk, you will learn about BrainGlobe, an initiative which exists to facilitate the development of interoperable Python-based tools for computational neuroanatomy.
This is the second of three lectures around current challenges and opportunities facing neuroinformatic infrastructure for handling sensitive data.
This lesson provides an overview of how to conceptualize, design, implement, and maintain neuroscientific pipelines in via the cloud-based computational reproducibility platform Code Ocean.