Skip to main content

This tutorial is part 1 of 2. It aims to provide viewers with an understanding of the fundamentals of R tool. Note: parts 1 and 2 of this tutorial are part of the same YouTube video; part 1 ends at 17:42. 

Difficulty level: Beginner
Duration: 17:42
Speaker: : Edureka

This lesson introduces the practical usage of The Virtual Brain (TVB) in its graphical user interface and via python scripts. In the graphical user interface, you are guided through its data repository, simulator, phase plane exploration tool, connectivity editor, stimulus generator, and the provided analyses. The implemented iPython notebooks of TVB are presented, and since they are public, can be used for further exploration of TVB. 

Difficulty level: Beginner
Duration: 1:12:24
Speaker: : Paul Triebkorn

This lesson provides a brief overview of the Python programming language, with an emphasis on tools relevant to data scientists.

Difficulty level: Beginner
Duration: 1:16:36
Speaker: : Tal Yarkoni

This tutorial covers the fundamentals of collaborating with Git and GitHub.

Difficulty level: Intermediate
Duration: 2:15:50
Speaker: : Elizabeth DuPre
Course:

This lesson provides a comprehensive introduction to the command line and 50 popular Linux commands. This is a long introduction (nearly 5 hours), but well worth it if you are going to spend a good part of your career working from a terminal, which is likely if you are interested in flexibility, power, and reproducibility in neuroscience research. This lesson is courtesy of freeCodeCamp.

Difficulty level: Beginner
Duration: 5:00:16
Speaker: : Colt Steele
Course:

This book was written with the goal of introducing researchers and students in a variety of research fields to the intersection of data science and neuroimaging. This book reflects our own experience of doing research at the intersection of data science and neuroimaging and it is based on our experience working with students and collaborators who come from a variety of backgrounds and have a variety of reasons for wanting to use data science approaches in their work. The tools and ideas that we chose to write about are all tools and ideas that we have used in some way in our own research. Many of them are tools that we use on a daily basis in our work. This was important to us for a few reasons: the first is that we want to teach people things that we ourselves find useful. Second, it allowed us to write the book with a focus on solving specific analysis tasks. For example, in many of the chapters you will see that we walk you through ideas while implementing them in code, and with data. We believe that this is a good way to learn about data analysis, because it provides a connecting thread from scientific questions through the data and its representation to implementing specific answers to these questions. Finally, we find these ideas compelling and fruitful. That’s why we were drawn to them in the first place. We hope that our enthusiasm about the ideas and tools described in this book will be infectious enough to convince the readers of their value.

 

Difficulty level: Intermediate
Duration:
Speaker: :
Course:

The Mouse Phenome Database (MPD) provides access to primary experimental trait data, genotypic variation, protocols and analysis tools for mouse genetic studies. Data are contributed by investigators worldwide and represent a broad scope of phenotyping endpoints and disease-related traits in naïve mice and those exposed to drugs, environmental agents or other treatments. MPD ensures rigorous curation of phenotype data and supporting documentation using relevant ontologies and controlled vocabularies. As a repository of curated and integrated data, MPD provides a means to access/re-use baseline data, as well as allows users to identify sensitized backgrounds for making new mouse models with genome editing technologies, analyze trait co-inheritance, benchmark assays in their own laboratories, and many other research applications. MPD’s primary source of funding is NIDA. For this reason, a majority of MPD data is neuro- and behavior-related.

Difficulty level: Beginner
Duration: 55:36
Speaker: : Elissa Chesler

This lecture covers the rationale for developing the DAQCORD, a framework for the design, documentation, and reporting of data curation methods in order to advance the scientific rigour, reproducibility, and analysis of data.

Difficulty level: Intermediate
Duration: 17:08
Speaker: : Ari Ercole

This is a tutorial on designing a Bayesian inference model to map belief trajectories, with emphasis on gaining familiarity with Hierarchical Gaussian Filters (HGFs).

 

This lesson corresponds to slides 65-90 of the PDF below. 

Difficulty level: Intermediate
Duration: 1:15:04
Speaker: : Daniel Hauke
Course:

An introduction to data management, manipulation, visualization, and analysis for neuroscience. Students will learn scientific programming in Python, and use this to work with example data from areas such as cognitive-behavioral research, single-cell recording, EEG, and structural and functional MRI. Basic signal processing techniques including filtering are covered. The course includes a Jupyter Notebook and video tutorials.

 

Difficulty level: Beginner
Duration: 1:09:16
Speaker: : Aaron J. Newman

This is a hands-on tutorial on PLINK, the open source whole genome association analysis toolset. The aims of this tutorial are to teach users how to perform basic quality control on genetic datasets, as well as to identify and understand GWAS summary statistics. 

Difficulty level: Intermediate
Duration: 1:27:18
Speaker: : Dan Felsky

In this third and final hands-on tutorial from the Research Workflows for Collaborative Neuroscience workshop, you will learn about workflow orchestration using open source tools like DataJoint and Flyte. 

Difficulty level: Intermediate
Duration: 22:36
Speaker: : Daniel Xenes

In this hands-on session, you will learn how to explore and work with DataLad datasets, containers, and structures using Jupyter notebooks. 

Difficulty level: Beginner
Duration: 58:05

This video will document the process of uploading data into a brainlife project using ezBIDS.

Difficulty level: Beginner
Duration: 6:15
Speaker: :

This short video walks you through the steps of publishing a dataset on brainlife, an open-source, free and secure reproducible neuroscience analysis platform.

Difficulty level: Beginner
Duration: 1:18
Speaker: :

This video will document the process of visualizing the provenance of each step performed to generate a data object on brainlife.

Difficulty level: Beginner
Duration: 0:21
Speaker: :

This video will document the process of downloading and running the "reproduce.sh" script, which will automatically run all of the steps to generate a data object locally on a user's machine.

Difficulty level: Beginner
Duration: 3:44
Speaker: :

This video will document the process of creating a pipeline rule for batch processing on brainlife.

Difficulty level: Intermediate
Duration: 0:57
Speaker: :

This short video shows how a brainlife.io publication can be opened from the Data Deposition page of the journal Nature Scientific Data.

Difficulty level: Beginner
Duration: 2:25
Speaker: :

In this tutorial, you will learn the basic features of uploading and versioning your data within OpenNeuro.org.

Difficulty level: Beginner
Duration: 5:36
Speaker: : OpenNeuro