This lesson gives a primer to project management in a scientific context, with a particular neuroinformatic case study.
In this lesson, you will hear about the current challenges regarding data management, as well as policies and resources aimed to address them.
This lesson provides an overview of how to manage relationships in a research context, while highlighting the need for effective communication at various levels.
Manipulate the default connectome provided with TVB to see how structural lesions effect brain dynamics. In this hands-on session you will insert lesions into the connectome within the TVB graphical user interface (GUI). Afterwards, the modified connectome will be used for simulations and the resulting activity will be analysed using functional connectivity.
This lecture covers FAIR atlases, including their background and construction, as well as how they can be created in line with the FAIR principles.
This lesson gives a description of the BrainHealth Databank, a repository of many types of health-related data, whose aim is to accelerate research, improve care, and to help better understand and diagnose mental illness, as well as develop new treatments and prevention strategies.
This lesson corresponds to slides 46-78 of the PDF below.
This lesson provides an overview of how to conceptualize, design, implement, and maintain neuroscientific pipelines in via the cloud-based computational reproducibility platform Code Ocean.
This lesson provides an overview of how to construct computational pipelines for neurophysiological data using DataJoint.
This hands-on tutorial walks you through DataJoint platform, highlighting features and schema which can be used to build robost neuroscientific pipelines.
This lesson provides an introduction to the DataLad, a free and open source distributed data management system that keeps track of your data, creates structure, ensures reproducibility, supports collaboration, and integrates with widely used data infrastructure.
This lesson introduces several open science tools like Docker and Apptainer which can be used to develop portable and reproducible software environments.
This lecture provides a detailed description of how to incorporate HED annotation into your neuroimaging data pipeline.
This lecture covers a wide range of aspects regarding neuroinformatics and data governance, describing both their historical developments and current trajectories. Particular tools, platforms, and standards to make your research more FAIR are also discussed.
This talk highlights a set of platform technologies, software, and data collections that close and shorten the feedback cycle in research.
This talk covers the Neuroimaging Informatics Tools and Resources Clearinghouse (NITRC), a free one-stop-shop collaboratory for science researchers that need resources such as neuroimaging analysis software, publicly available data sets, or computing power.
This lecture covers how to make modeling workflows FAIR by working through a practical example, dissecting the steps within the workflow, and detailing the tools and resources used at each step.
This lecture focuses on the structured validation process within computational neuroscience, including the tools, services, and methods involved in simulation and analysis.
This lecture discusses the FAIR principles as they apply to electrophysiology data and metadata, the building blocks for community tools and standards, platforms and grassroots initiatives, and the challenges therein.
This session provides users with an introduction to tools and resources that facilitate the implementation of FAIR in their research.
This session will include presentations of infrastructure that embrace the FAIR principles developed by members of the INCF Community.
This lecture provides an overview of The Virtual Brain Simulation Platform.