This tutorial demonstrates how to perform cell-type deconvolution in order to estimate how proportions of cell-types in the brain change in response to various conditions. While these techniques may be useful in addressing a wide range of scientific questions, this tutorial will focus on the cellular changes associated with major depression (MDD).
In this lesson, you will learn about data management within the Open Data Commons (ODC) framework, and in particular, how Spinal Cord Injury (SCI) data is stored, shared, and published. You will also hear about Frictionless Data, an open-source toolkit aimed at simplifying the data experience.
This lesson introduces several open science tools like Docker and Apptainer which can be used to develop portable and reproducible software environments.
This talk covers the differences between applying HED annotation to fMRI datasets versus other neuroimaging practices, and also introduces an analysis pipeline using HED tags.
This lesson provides a brief visual walkthrough on the necessary steps when copying data from one brainlife project to another.
This lesson visually documents the process of uploading data to brainlife via the command line interface (CLI).
This video shows how to use the brainlife.io interface to edit the participants' info file. This file is the ParticipantInfo.json file of the Brain Imaging Data Structure (BIDS).
This video will document the process of running an app on brainlife, from data staging to archiving of the final data outputs.
This video demonstrates each required step for preprocessing T1w anatomical data in brainlife.io.
This short video shows how data in a brainlife.io publication can be opened from a DOI inside a published article. The video provides an example of how the DOI deposited on the journal can be opened with a web browser to redirect to the associated data publication on brainlife.io.
This lecture contains an overview of electrophysiology data reuse within the EBRAINS ecosystem.
This video explains what metadata is, why it is important, and how you can organize your metadata to increase the FAIRness of your data on EBRAINS.
This lecture covers the rationale for developing the DAQCORD, a framework for the design, documentation, and reporting of data curation methods in order to advance the scientific rigour, reproducibility, and analysis of data.
This is the first of two workshops on reproducibility in science, during which participants are introduced to concepts of FAIR and open science. After discussing the definition of and need for FAIR science, participants are walked through tutorials on installing and using Github and Docker, the powerful, open-source tools for versioning and publishing code and software, respectively.
In this lesson, while learning about the need for increased large-scale collaborative science that is transparent in nature, users also are given a tutorial on using Synapse for facilitating reusable and reproducible research.
This lesson contains the first part of the lecture Data Science and Reproducibility. You will learn about the development of data science and what the term currently encompasses, as well as how neuroscience and data science intersect.
In this second part of the lecture Data Science and Reproducibility, you will learn how to apply the awareness of the intersection between neuroscience and data science (discussed in part one) to an understanding of the current reproducibility crisis in biomedical science and neuroscience.
The lecture provides an overview of the core skills and practical solutions required to practice reproducible research.
This lecture provides an introduction to reproducibility issues within the fields of neuroimaging and fMRI, as well as an overview of tools and resources being developed to alleviate the problem.
This lecture provides a historical perspective on reproducibility in science, as well as the current limitations of neuroimaging studies to date. This lecture also lays out a case for the use of meta-analyses, outlining available resources to conduct such analyses.