This talk describes the NIH-funded SPARC Data Structure, and how this project navigates ontology development while keeping in mind the FAIR science principles.
This lecture covers structured data, databases, federating neuroscience-relevant databases, and ontologies.
This lecture covers FAIR atlases, including their background and construction, as well as how they can be created in line with the FAIR principles.
This lecture covers the NIDM data format within BIDS to make your datasets more searchable, and how to optimize your dataset searches.
This lecture covers positron emission tomography (PET) imaging and the Brain Imaging Data Structure (BIDS), and how they work together within the PET-BIDS standard to make neuroscience more open and FAIR.
This lecture discusses how to standardize electrophysiology data organization to move towards being more FAIR.
Hierarchical Event Descriptors (HED) fill a major gap in the neuroinformatics standards toolkit, namely the specification of the nature(s) of events and time-limited conditions recorded as having occurred during time series recordings (EEG, MEG, iEEG, fMRI, etc.). Here, the HED Working Group presents an online INCF workshop on the need for, structure of, tools for, and use of HED annotation to prepare neuroimaging time series data for storing, sharing, and advanced analysis.
In this lesson, attendees will learn about the data structure standards, specifically the Brain Imaging Data Structure (BIDS), an INCF-endorsed standard for organizing, annotating, and describing data collected during neuroimaging experiments.
This lecture contains an overview of the Distributed Archives for Neurophysiology Data Integration (DANDI) archive, its ties to FAIR and open-source, integrations with other programs, and upcoming features.
This lesson gives an introduction to the Mathematics chapter of Datalabcc's Foundations in Data Science series.
This lesson serves a primer on elementary algebra.
This lesson provides a primer on linear algebra, aiming to demonstrate how such operations are fundamental to many data science.
In this lesson, users will learn about linear equation systems, as well as follow along some practical use cases.
This talk gives a primer on calculus, emphasizing its role in data science.
This lesson clarifies how calculus relates to optimization in a data science context.
This lesson covers Big O notation, a mathematical notation that describes the limiting behavior of a function as it tends towards a certain value or infinity, proving useful for data scientists who want to evaluate their algorithms' efficiency.
This lesson serves as a primer on the fundamental concepts underlying probability.
Serving as good refresher, this lesson explains the maths and logic concepts that are important for programmers to understand, including sets, propositional logic, conditional statements, and more.
This compilation is courtesy of freeCodeCamp.
This lesson provides a useful refresher which will facilitate the use of Matlab, Octave, and various matrix-manipulation and machine-learning software.
This lesson was created by RootMath.
This lecture covers a lot of post-war developments in the science of the mind, focusing first on the cognitive revolution, and concluding with living machines.