This lecture will provide an overview of the INCF Training Suite, a collection of tools that embraces the FAIR principles developed by members of the INCF Community. This will include an overview of TrainingSpace, Neurostars, and KnowledgeSpace.
This session will include presentations of infrastructure that embrace the FAIR principles developed by members of the INCF Community. This lecture provides an overview and demo of the Canadian Open Neuroscience Platform (CONP).
This lecture contains an overview of the China-Cuba-Canada neuroinformatics ecosystem for Quantitative Tomographic EEG Analysis (qEEGt).
This session provides users with an introduction to tools and resources that facilitate the implementation of FAIR in their research.
This lesson provides an overview of self-supervision as it relates to neural data tasks and the Mine Your Own vieW (MYOW) approach.
In this talk, you will learn how brainlife.io works, and how it can be applied to neuroscience data.
This lecture covers the IBI Data Standards and Sharing Working Group, including its history, aims, and projects.
This session covers the framework of the International Brain Lab (IBL) and the data architecture used for this project.
This session will include presentations of infrastructure that embrace the FAIR principles developed by members of the INCF Community.
This lecture provides an overview of The Virtual Brain Simulation Platform.
This lecture provides an introduction to the course "Cognitive Science & Psychology: Mind, Brain, and Behavior".
This lesson covers the history of neuroscience and machine learning, and the story of how these two seemingly disparate fields are increasingly merging.
In this lesson you will learn how machine learners and neuroscientists construct abstract computational models based on various neurophysiological signalling properties.
In this lesson, you will learn about some typical neuronal models employed by machine learners and computational neuroscientists, meant to imitate the biophysical properties of real neurons.
Whereas the previous two lessons described the biophysical and signalling properties of individual neurons, this lesson describes properties of those units when part of larger networks.
This lesson goes over some examples of how machine learners and computational neuroscientists go about designing and building neural network models inspired by biological brain systems.
In this lesson, you will learn about different approaches to modeling learning in neural networks, particularly focusing on system parameters such as firing rates and synaptic weights impact a network.
In this lesson, you will learn more about some of the issues inherent in modeling neural spikes, approaches to ameliorate these problems, and the pros and cons of these approaches.
In this lesson, you will learn about some of the many methods to train spiking neural networks (SNNs) with either no attempt to use gradients, or only use gradients in a limited or constrained way.
In this lesson, you will learn how to train spiking neural networks (SNNs) with a surrogate gradient method.
This lesson explores how researchers try to understand neural networks, particularly in the case of observing neural activity.