This lecture covers FAIR atlases, including their background and construction, as well as how they can be created in line with the FAIR principles.
This lecture covers the NIDM data format within BIDS to make your datasets more searchable, and how to optimize your dataset searches.
This lecture covers positron emission tomography (PET) imaging and the Brain Imaging Data Structure (BIDS), and how they work together within the PET-BIDS standard to make neuroscience more open and FAIR.
This lecture discusses the FAIR principles as they apply to electrophysiology data and metadata, the building blocks for community tools and standards, platforms and grassroots initiatives, and the challenges therein.
This lecture contains an overview of electrophysiology data reuse within the EBRAINS ecosystem.
This lecture contains an overview of the Distributed Archives for Neurophysiology Data Integration (DANDI) archive, its ties to FAIR and open-source, integrations with other programs, and upcoming features.
This lecture contains an overview of the Australian Electrophysiology Data Analytics Platform (AEDAPT), how it works, how to scale it, and how it fits into the FAIR ecosystem.
This lecture discusses how to standardize electrophysiology data organization to move towards being more FAIR.
This lecture will provide an overview of the INCF Training Suite, a collection of tools that embraces the FAIR principles developed by members of the INCF Community. This will include an overview of TrainingSpace, Neurostars, and KnowledgeSpace.
This lecture contains an overview of the China-Cuba-Canada neuroinformatics ecosystem for Quantitative Tomographic EEG Analysis (qEEGt).
This module explains how neurons come together to create the networks that give rise to our thoughts. The totality of our neurons and their connection is called our connectome. Learn how this connectome changes as we learn, and computes information.
This session discussed the secret life of your dataset metadata: the ways in which, for many years to come, it will work non-stop to foster the visibility, reach, and impact of your work. We explored how metadata will help your dataset travel through the global research infrastructure, and how data repositories and discovery services can use this metadata to help launch your dataset into the world.
This lesson provides information on developing data management plans (DMPs), including an overview of how DMPs contribute to effective research efforts, as well as specific development resources and DMP examples.
In this session, participants will take an in-depth look at the newly launched DMP Assistant 2.0, including all of its enhanced key features for both end-users and institutional administrators, as well as a brief look at the future of the platform.
This lecture provides reviews some standards for project management and organization, including motivation from the view of the FAIR principles and improved reproducibility.
This lesson provides a conceptual overview of the rudiments of machine learning, including its bases in traditional statistics and the types of questions it might be applied to. The lesson was presented in the context of the BrainHack School 2020.
This lesson provides a hands-on, Jupyter-notebook-based tutorial to apply machine learning in Python to brain-imaging data.
This lesson presents advanced machine learning algorithms for neuroimaging, while addressing some real-world considerations related to data size and type.
This lesson gives an introduction to high-performance computing with the Compute Canada network, first providing an overview of use cases for HPC and then a hands-on tutorial. Though some examples might seem specific to the Calcul Québec, all computing clusters in the Compute Canada network share the same software modules and environments.
This lesson provides a short overview of the main features of the Canadian Open Neuroscience Platform (CONP) Portal, a web interface that facilitates open science for the neuroscience community by simplifying global access to and sharing of datasets and tools. The Portal internalizes the typical cycle of a research project, beginning with data acquisition, followed by data processing with published tools, and ultimately the publication of results with a link to the original dataset.