This lesson continues with the second workshop on reproducible science, focusing on additional open source tools for researchers and data scientists, such as the R programming language for data science, as well as associated tools like RStudio and R Markdown. Additionally, users are introduced to Python and iPython notebooks, Google Colab, and are given hands-on tutorials on how to create a Binder environment, as well as various containers in Docker and Singularity.
This talk describes the relevance and power of using brain atlases as part of one's data integration pipeline.
In this lesson, you will learn how to understand data management plans and why data sharing is important.
This quick visual walkthrough presents the steps required in uploading data into a brainlife project using the graphical user interface (GUI).
This short walkthrough documents the steps needed to find a dataset in OpenNeuro, a free and open platform for sharing MRI, MEG, EEG, iEEG, ECoG, ASL, and PET data, and import it directly to a brainlife project.
This lesson describes and shows four different ways one may upload their data to brainlife.io.
This lecture covers why data sharing and other collaborative practices are important, how these practices are developed, and the challenges involved in their development and implementation.
This lecture discusses the FAIR principles as they apply to electrophysiology data and metadata, the building blocks for community tools and standards, platforms and grassroots initiatives, and the challenges therein.
This lecture contains an overview of electrophysiology data reuse within the EBRAINS ecosystem.
This lecture contains an overview of the Distributed Archives for Neurophysiology Data Integration (DANDI) archive, its ties to FAIR and open-source, integrations with other programs, and upcoming features.
This lesson provides a short overview of the main features of the Canadian Open Neuroscience Platform (CONP) Portal, a web interface that facilitates open science for the neuroscience community by simplifying global access to and sharing of datasets and tools. The Portal internalizes the typical cycle of a research project, beginning with data acquisition, followed by data processing with published tools, and ultimately the publication of results with a link to the original dataset.
This lesson gives an introduction to the Mathematics chapter of Datalabcc's Foundations in Data Science series.
This lesson serves a primer on elementary algebra.
This lesson provides a primer on linear algebra, aiming to demonstrate how such operations are fundamental to many data science.
In this lesson, users will learn about linear equation systems, as well as follow along some practical use cases.
This talk gives a primer on calculus, emphasizing its role in data science.
This lesson clarifies how calculus relates to optimization in a data science context.
This lesson covers Big O notation, a mathematical notation that describes the limiting behavior of a function as it tends towards a certain value or infinity, proving useful for data scientists who want to evaluate their algorithms' efficiency.
This lesson serves as a primer on the fundamental concepts underlying probability.
Serving as good refresher, this lesson explains the maths and logic concepts that are important for programmers to understand, including sets, propositional logic, conditional statements, and more.
This compilation is courtesy of freeCodeCamp.